Skip to main content
Version: 23.2.0

Google Life Sciences

This guide assumes you have an existing Google Cloud account. Sign-up for a free account here.

Tower provides integration to Google Cloud via the Cloud Life Sciences API.

This guide is split into two parts:

  1. How to configure your Google Cloud account to use the Cloud Life Sciences API.

  2. How to create a Google Life Sciences compute environment in Tower.

Configure Google Cloud

Create a project

Navigate to the Google Project Selector page and either select an existing project or select Create project.

Enter a name for your new project, e.g., "tower-nf".

If you are part of an organization, the location will default to your organization.

Enable billing

See here to enable billing in your Google Cloud account.

Enable APIs

See here to enable the following APIs for your project:

  • Cloud Life Sciences API
  • Compute Engine API
  • Cloud Storage API

Select your project from the dropdown menu and select Enable.

Alternatively, select your project in the navigation bar and enable each API manually from these pages:

Create a service account key

  1. In the navigation menu, select IAM & Admin, then Service Accounts.

  2. Select the email address of the Compute Engine default service account.

  3. Select Keys, then Add key, then Create new key.

  4. Select JSON as the key type.

  5. Select Create.

A JSON file will be downloaded to your computer. This file contains the credential needed to configure the compute environment in Tower.

You can manage your key from the Service Accounts page.

Create a Cloud Storage bucket

  1. In the navigation menu (), select Cloud Storage, then Create bucket.

  2. Enter a name for your bucket. You will reference this name when creating the compute environment in Tower.

Do not use underscores (_) in your bucket name. Use hyphens (-) instead.

  1. Select Region for the Location type and select the Location for your bucket. You will reference this location when creating the compute environment in Tower.

  2. Select Standard for the default storage class.

  3. Select Uniform for the Access control.

The Cloud Life Sciences API is available in a limited number of locations. These locations are only used to store metadata about the pipeline operations. The storage bucket and compute resources can be in any region.

  1. Select Create.

  2. Once the bucket is created, you will be redirected to the Bucket details page.

  3. Select Permissions, then + Add.

  4. Copy the email address of the Compute Engine default service account into New principals.

  5. Select the following roles:

  • Storage Admin
  • Storage Legacy Bucket Owner
  • Storage Legacy Object Owner
  • Storage Object Creator

Compute environment

The following guide to configure Tower assumes you have (1) a service account key for a Google Cloud account and (2) the name and location of a Cloud Storage bucket.

To create a new compute environment for Google Cloud in Tower:

  1. In a workspace, select Compute Environments and then New Environment.

  2. Enter a descriptive name for this environment, e.g., "Google Life Sciences (europe-west2)".

  3. Select Google Life Sciences as the target platform.

  4. From the Credentials drop-down, select existing Google Cloud credentials, or add new credentials by selecting the + button. If you select to use existing credentials, skip to step 7.

  5. Enter a name for the credentials, e.g. "Google Cloud Credentials".

  6. Enter the Service account key created previously.

You can create multiple credentials in your Tower workspace.

  1. Select the Region and Zones where you wish to execute pipelines. Leave the Location empty for the Cloud Life Sciences API to use the closest available location.

  2. In the Pipeline work directory field, enter your storage bucket URL, e.g., gs://my-bucket. This bucket should be accessible in the region selected in the previous step.

  3. You can enable Preemptible to use preemptible instances, which have significantly reduced cost compared to on-demand instances.

  4. You can use a Filestore file system to automatically mount a Google Filestore volume in your pipelines.

  5. Apply Resource labels to the cloud resources consumed by this compute environment. Workspace default resource labels are prefilled.

  6. Expand Staging options to include optional pre- or post-run Bash scripts that execute before or after the Nextflow pipeline execution in your environment.

  7. Use the Environment variables option to specify custom environment variables for the Head job and/or Compute jobs.

  8. Configure any advanced options described below, as needed.

  9. Select Create to finalize the compute environment setup.

Jump to the documentation for launching pipelines.

Advanced options

  • Enable Use Private Address to ensure that your Google Cloud VMs aren't accessible to the public internet.

  • Use Boot disk size to control the boot disk size of VMs.

  • Use Head Job CPUs and Head Job Memory to specify the CPUs and memory allocated for head jobs.