The project pipeline
This section covers how to develop, run, and test your code to ensure it will work end-to-end within the secure framework.
Project pipelines🔗
The cohortextractor section describes how to make an action which generate dummy datasets based on the instructions defined in your study_definition.py
script.
These dummy datasets are the basis for developing the analysis code that will eventually be passed to the server to run on real datasets.
The code can be written and run on your local machine using whatever development set up you prefer (e.g., developing R in RStudio).
However, it's important to ensure that this code will run successfully in OpenSAFELY's secure environment too, using the specific language and package versions that are installed there. To do this, you should use the project pipeline.
The project pipeline, defined entirely in a project.yaml
file, is a system for executing your code using a series of actions i.e., a discrete analytical step within the analysis, each of which may depend on previous actions.
The primary purpose of the pipeline is to specify the execution order for all your code, so that it can be automatically run and tested from start to finish using dummy data and using the live database in the secure environment, using an identical software configuration. Arranging your code like this also has several other advantages:
- The pipeline knows if outputs for given actions already exist, and by default skips running them if so. This greatly speeds up the debugging cycle when testing against live data
- In production, actions that can be executed in parallel will be, automatically
- Thinking about your analysis in terms of actions makes it more readable and therefore easier to review and test. For example, being explicit about what the inputs and outputs of each actions are ensures you don't overwrite files by accident.
- The pipeline forces you to declare which outputs may be more or less disclosive.
project.yaml
format🔗
The project pipeline is defined in a single file, project.yaml
, which lives in the repository's root directory.
It is written using a configuration format called YAML, which uses indentation to indicate groupings of related variables.
A simple example of a project.yaml
is as follows:
version: "3.0"
expectations:
population_size: 1000
actions:
generate_study_population:
run: cohortextractor:latest generate_cohort --study-definition study_definition
outputs:
highly_sensitive:
cohort: output/input.csv
run_model:
run: stata-mp:latest analysis/model.do
needs: [generate_study_population]
outputs:
moderately_sensitive:
model: models/cox-model.txt
figure: figures/survival-plot.png
This example declares the pipeline version
, the population_size
for the dummy data, and two actions, generate_study_population
and run_model
.
You only need to change version
if you want to take advantage of features of newer versions of the pipeline framework.
The generate_study_population
action will create the highly sensitive input.csv
dataset.
It will be dummy data when run locally, and will be based on real data from the OpenSAFELY database when run in the secure environment.
The run_model
action will run a Stata script called model.do
based on the the input.csv
created by the previous action.
It will output two moderately sensitive files cox-model.txt
and survival-plot.png
, which can be checked and released if appropriate.
Every project.yaml
requires a version
, expectations
, and actions
section.
In general, actions are composed as follows:
- Each action must be named using a valid YAML key (you won't go wrong with letters, numbers, and underscores) and must be unique.
- Each action must include a
run
key which includes an officially-supported command and a version (latest
will always select the most recent version, but following initial development you should specify the version to ensure reproducibility).- The
cohortextractor
command has the same options as described in the cohortextractor section. - The
python
,r
, andstata-mp
commands provide a locked-down execution environment that can take one or moreinputs
which are passed to the code.
- The
- Each action must include an
outputs
key with at least one output, classified as eitherhighly_sensitive
ormoderately_sensitive
highly_sensitive
outputs are considered potentially highly-disclosive, and are never intended for publishing outside the secure environmentmoderately_sensitive
outputs are automatically copied to the secure review area for redaction (otherwise known as Level 4) and potentially for publication back to GitHub.
- Each action can include a
needs
key which specifies a list of actions (contained within square brackets and separated by commas) that are required for it to successfully run. When an action runs, theoutputs
of all itsneeds
actions are copied to its working directory.needs
actions can be defined anywhere in theproject.yaml
, but it's more readable if they are defined above.
When writing and running your pipeline, note that:
-
All file paths must be declared relative to the repository's root directory. So for example use
outputs/figures/
, notC:/users/elvis/documents/myrepo/outputs/figures
. -
File paths are case-sensitive as everything is run inside a Linux Docker container.
-
The location of each action's output is determined by the underlying code that the action invoked, not by the value of the
outputs
configuration. The purpose ofoutputs
is to label the disclosivity of each output and indicate that it should be stored securely — any outputs not labelled will not be saved. -
Each action is run in its own isolated environment in a temporary working directory. This means that all the necessary libraries and data must be imported within the script for each action — For R users, this essentially means that the R is restarted for each action.
-
If one or more dependencies of an action have not been run (i.e., their outputs do not exist) then these dependency actions will be run first. If a dependency has changed but has not been run (so the outputs are not up-to-date with the changes), then the dependency actions will not be run, and the dependent actions will be run using the out-of-date outputs.
Running your code locally🔗
Whilst you can develop and run code locally using your own installations of R, Stata or Python, it's important to check that these will also successfully run on the real data in an identical execution environment.
The opensafely run
command will execute one or more actions according to the project.yaml
.
To see its options, type opensafely run --help
.
For opensafely run
to work:
- You need to have both Python and Docker installed.
- The Docker daemon must be running on your machine:
- For Windows users using Docker Desktop, there should be a Docker icon in your system tray.
- For Mac users using Docker Desktop, there should be a Docker icon in the top status bar.
To run the first action in the example above, using dummy data, you can use:
opensafely run generate_study_population
This will generate the input.csv
file as explained in the cohortextractor section.
To run the second action you can use:
opensafely run run_model
It will create the two files as specified in the analysis/model.do
script.
To force the dependencies to be run you can use for example opensafely run run_model --force-run-dependencies
, or -f
for short.
This will ensure for example that both the run_model
and generate_study_population
actions are run, even if input.csv
already exists.
To run all actions, you can use a special run_all
action which is created for you (no need to define it in your project.yaml
):
opensafely run run_all
Each time an action is run, logging information about your run will be put into the metadata/
folder.
If any of your actions fail, you may find clues here as to why.
Click here for information on the exact steps that occur when each job is run locally
What happens:
- A new, empty temporary directory for the job is created
- Any files in the local repo that do not match the output patterns in the
project.yaml
are copied into the temporary folder - Any output files from the job's dependencies are copied into the temporary folder
- The job is run
- All the files matching the specified output patterns are copied into the local repo
- The log files for the job are saved into the
metadata/
directory - The temporary directory is deleted
Running your code with GitHub actions🔗
Every time you create a pull request to merge a development branch onto the main remote branch, GitHub will automatically run a series of tests on the code; specifically, that your codelists are up-to-date, and that run_all
completes successfully.
Depending on your settings, you may receive email notifications about the results of these tests.
You can view the tests, including any errors or failures, by going to the pull request page on GitHub and clicking the checks
tab.
You can re-run these tests by clicking the re-run jobs
button.
Running your code on the server🔗
To run code for real in the production environment, use the job server website.
Accessing outputs🔗
After your project has been executed via the job server, its outputs will be stored on a secure server.
Users with permission to access Level 4 can view output files that are labelled as moderately sensitive; they can also view automatically created log files of the run for debugging purposes.
For security reasons, they will be in a different directory than if you had run locally. For the TPP backend, outputs labelled moderately_sensitive
in the project.yaml
will be saved in D:/Level4Files/workspaces/<NAME_OF_YOUR_WORKSPACE>
. These outputs can be reviewed on the server and released via GitHub if they are deemed non-disclosive.
Outputs labelled highly_sensitive
are not visible.
If you have Level 3 access🔗
No data should ever be published from the Level 3 server. Access is only for permitted users, for the purpose of debugging problems in the secure environment.
Highly sensitive outputs can be seen in E:/high_privacy/workspaces/<WORKSPACE_NAME>
. This includes a directory called metadata
, containing log files for each action e.g. generate_cohorts.log
, run_model.log
.
Moderately sensitive outputs can be seen in E:/FILESFORL4/workspaces/<WORKSPACE_NAME>
.
Running your code manually in the server🔗
This is only possible for people with Level 3 access. You'll want to refer to instructions for interacting with OpenSAFELY via the secure server (in restricted access repo).
The live environment is set up via a wrapper script; instead of cohortextractor
, you should run /e/bin/actionrunner.sh
.
For example, to run run_model
on the Level 3 server, against the full
database, you'd type:
/e/bin/actionrunner.sh run full run_model tpp