Skip to main content

Fusion using local execution and S3 object storage

Fusion allow running Nextflow pipeline using the local executor and a S3 bucket as the pipeline scratch directory. This can be useful to scale vertically your pipeline execution using a large EC2 instance and without requiring to allocate a large EBS volume for the pipeline temporary data.

The pipeline configuration looks like the following:

wave.enabled = true
docker.enabled = true
fusion.enabled = true
fusion.exportAwsAccessKeys = true

Save the above snippet to a file named nextflow.config into the pipeline launching directory.

Then launch the pipeline execution with the usual run command:

nextflow run <YOUR PIPELINE SCRIPT> -w s3://<YOUR-BUCKET>/work

Replacing YOUR PIPELINE SCRIPT with the URI of your pipeline Git repository and YOUR-BUCKET with a S3 bucket of your choice.

To achieve best performance make sure to setup a SSD volumes as temporary directory. See the section SSD storage for details.