WDL is a workflow language meant to be read and written by humans. Broader documentation is provided by the WDL website. Any questions or issues can be discussed at our support forum.
Library and engine support is provided by
- Java parser which provides only a parser to convert a WDL string into an AST
- wdl4s provides Scala bindings for WDL and uses the above Java parser
- PyWDL provides Python bindings for WDL
- Cromwell is an engine for running WDL workflows. This uses wdl4s
The Workflow Description Language is a domain specific language for describing tasks and workflows.
An example WDL file that describes three tasks to run UNIX commands (in this case, ps
, grep
, and wc
) and then link them together in a workflow would look like this:
task ps {
command {
ps
}
output {
File procs = stdout()
}
}
task cgrep {
String pattern
File in_file
command {
grep '${pattern}' ${in_file} | wc -l
}
output {
Int count = read_int(stdout())
}
}
task wc {
File in_file
command {
cat ${in_file} | wc -l
}
output {
Int count = read_int(stdout())
}
}
workflow three_step {
call ps
call cgrep {
input: in_file=ps.procs
}
call wc {
input: in_file=ps.procs
}
}
WDL aims to be able to describe tasks with abstract commands which have inputs. Abstract commands are a template with parts of the command left for the user to provide a value for. In the example above, the task wc
declaration defines a task with one input (in_file
of type file) and one output (count
of type int).
Once tasks are defined, WDL allows you to construct a workflow of these tasks. Since each task defines its inputs and outputs explicitly, you can wire together one task's output to be another task's input and create a dependency graph. An execution engine can then collect the set of inputs it needs from the user to run each task in the workflow up front and then run the tasks in the right order.
WDL also lets you define more advanced structures, like the ability to call a task in parallel (referred to as 'scattering'). In the example below, the wc
task is being called n-times where n is the length of the Array[String] str_array
variable. Each element of the str_array
is used as the value of the str
parameter in the call to the wc
task.
task wc {
String str
command {
echo "${str}" | wc -c
}
output {
Int count = read_int(stdout()) - 1
}
}
workflow wf {
Array[String] str_array
scatter(s in str_array) {
call wc {
input: str=s
}
}
}
We'll use Cromwell and wdltool to run these examples but you can use any WDL engine of your choice.
If you don't already have a reference to the Cromwell JAR file, one can be downloaded
If you don't already have a reference to the wdltool JAR file, one can be downloaded
Create a WDL simple file and save it as hello.wdl
, for example:
task hello {
String name
command {
echo 'Hello ${name}!'
}
output {
File response = stdout()
}
}
workflow test {
call hello
}
Generate a template hello.json
file with the inputs
subcommand:
$ java -jar wdltool.jar inputs hello.wdl
{
"test.hello.name": "String"
}
WDL has a concept of fully-qualified names. In the above output, test.hello.name
is a fully-qualified name which should be read as: the name
input on the hello
call within workflow test
. Fully-qualified names are used to unambiguously refer to specific elements of a workflow. All inputs are specified by fully-qualified names and all outputs are returned as fully-qualified names.
Modify this and save it to hello.json
:
{
"test.hello.name": "World"
}
Then, use the run
subcommand to run the workflow:
$ java -jar cromwell.jar run hello.wdl hello.json
... truncated ...
{
"test.hello.response": "/home/user/test/c1d15098-bb57-4a0e-bc52-3a8887f7b439/call-hello/stdout8818073565713629828.tmp"
}
Since the hello
task returns a File
, the result is a file that contains the string "Hello World!" in it.
$ cat /home/user/test/c1d15098-bb57-4a0e-bc52-3a8887f7b439/call-hello/stdout8818073565713629828.tmp
hello world!
Currently the hello
task returns a File
with the greeting in it, but what if we wanted to return a String
instead?
This can be done by utilizing the read_string()
function:
task hello {
String name
command {
echo 'Hello ${name}!'
}
output {
String response = read_string(stdout())
}
}
workflow test {
call hello
}
Now when this is run, we get the string output for test.hello.response
:
$ java -jar cromwell.jar run hello.wdl hello.json
... truncated ...
{
"test.hello.response": "Hello World!"
}
read_string
is a function in the standard library, which provides other useful functions for converting outputs to WDL data types.
So far we've only been dealing with the standard output of a command, but what if it writes a file to disk? Consider this example:
task hello {
String name
command {
echo 'Hello ${name}!' > test.out
}
output {
String response = read_string("test.out")
}
}
workflow test {
call hello
}
Now when this is run, we get the string output for test.hello.response
:
$ java -jar cromwell.jar run hello.wdl hello.json
... truncated ...
{
"test.hello.response": "Hello World!"
}
read_string
is a function in the standard library, which provides other useful functions for converting outputs to WDL data types.
We can use the glob() function to read multiple files at once:
task globber {
command <<<
for i in `seq 1 5`
do
mkdir out-$i
echo "globbing is my number $i best hobby" > out-$i/$i.txt
done
>>>
output {
Array[File] outFiles = glob("out-*/*.txt")
}
}
workflow test {
call globber
}
Now when this is run, the outFiles
output array will contain all files
found by evaluating the specified glob.
$ java -jar cromwell.jar run globber.wdl -
... truncated ...
{
"test.globber.outFiles": ["/home/user/test/dee60566-267b-4f33-a1dd-0b199e6292b8/call-globber/out-3/3.txt", "/home/user/test/dee60566-267b-4f33-a1dd-0b199e6292b8/call-globber/out-5/5.txt", "/home/user/test/dee60566-267b-4f33-a1dd-0b199e6292b8/call-globber/out-2/2.txt", "/home/user/test/dee60566-267b-4f33-a1dd-0b199e6292b8/call-globber/out-4/4.txt", "/home/user/test/dee60566-267b-4f33-a1dd-0b199e6292b8/call-globber/out-1/1.txt"]
}
This workflow has no inputs, so a "-" was passed to Cromwell for the inputs file.
Sometimes, an output file is named as a function of one of its inputs.
task hello {
String name
command {
echo 'Hello ${name}!' > ${name}.txt
}
output {
String response = read_string("${name}.txt")
}
}
workflow test {
call hello
}
Here the inputs and outputs are exactly the same as previous examples, however the intermediate output file name of this task is named differently for every invocation.
Say we wanted to call the hello
task twice. Simply adding two call hello
statements to the body of workflow test
would result in non-unique fully-qualified names. To resolve this issue, call
statements can be aliased using an as
clause:
task hello {
String name
command {
echo 'Hello ${name}!'
}
output {
String response = read_string(stdout())
}
}
workflow test {
call hello
call hello as hello2
}
Now, we need to specify a value for test.hello2.name
in the hello.json file:
{
"test.hello.name": "World",
"test.hello2.name": "Boston"
}
Running this workflow now produces two outputs:
$ java -jar cromwell.jar run hello.wdl hello.json
... truncated ...
{
"test.hello.response": "Hello World!",
"test.hello2.response": "Hello Boston!"
}
A call
can have an optional section to define inputs. As seen below, the key/value pairs represent the name of the input on the left-hand side and the expression for the input's value on the right-hand side:
task hello {
String name
String salutation
command {
echo '${salutation} ${name}!'
}
output {
String response = read_string(stdout())
}
}
workflow test {
call hello {
input: salutation="Greetings"
}
call hello as hello2
}
Now, the hello.json
would require three inputs:
{
"test.hello.name": "World",
"test.hello2.name": "Boston",
"test.hello2.salutation": "Hello"
}
Running this workflow still gives us the two greetings we expect:
$ java -jar cromwell.jar run hello.wdl hello.json
... truncated ...
{
"test.hello.response": "Greetings World!",
"test.hello2.response": "Hello Boston!"
}
What if we wanted to parameterize the greeting and make it used for all invocations of task hello
? In this situation, a declaration can be used:
task hello {
String salutation
String name
command {
echo '${salutation}, ${name}!'
}
output {
String response = read_string(stdout())
}
}
workflow test {
String greeting
call hello {
input: salutation=greeting
}
call hello as hello2 {
input: salutation=greeting + " and nice to meet you"
}
}
String greeting
is referenced to satisfy the "salutation" parameter to both invocations of the hello
task.
The inputs required to run this would be:
{
"test.hello.name": "World",
"test.hello2.name": "Boston",
"test.greeting": "Hello"
}
And this would produce the following outputs when run
$ java -jar cromwell.jar run hello.wdl hello.json
... truncated ...
{
"test.hello.response": "Hello, World!",
"test.hello2.response": "Hello and nice to meet you, Boston!"
}
So far every example has used the default type of String
for every input. Passing files along to tasks is simply a matter of defining the input type as File
:
task grep {
File file
command {
grep -c '^...$' ${file}
}
output {
Int count = read_int(stdout())
}
}
workflow test {
call grep
}
The read_int()
function here would read the contents of its parameter, and interpret the first line as an integer and return that value as a WDL Int
type.
If I specified a file called test_file
with the contents of:
foo
bar
baz
quux
And then the inputs JSON file would be:
{
"test.grep.file": "test_file"
}
The result of running this would be:
$ java -jar cromwell.jar run grep.wdl grep.json
... truncated ...
{
"test.grep.count": 3
}
Scatter blocks can be used to run the same call multiple times but only varying a specific parameter on each invocation. Consider the following example:
task prepare {
command <<<
python -c "print('one\ntwo\nthree\nfour')"
>>>
output {
Array[String] array = read_lines(stdout())
}
}
task analysis {
String str
command <<<
python -c "print('_${str}_')"
>>>
output {
String out = read_string(stdout())
}
}
task gather {
Array[String] array
command <<<
echo ${sep=' ' array}
>>>
output {
String str = read_string(stdout())
}
}
workflow example {
call prepare
scatter (x in prepare.array) {
call analysis {input: str=x}
}
call gather {input: array=analysis.out}
}
This example calls the analysis
task once for each element in the array that the prepare
task outputs. The resulting outputs of this workflow would be:
$ java -jar cromwell.jar run scatter.wdl -
... truncated ...
{
"example.analysis.out": ["_one_", "_two_", "_three_", "_four_"],
"example.gather.str": "_one_ _two_ _three_ _four_",
"example.prepare.array": ["one", "two", "three", "four"]
}