Import data from Gen3 BDC and Gen3 AnVIL

Overview

The purpose of this feature is to allow you bring data to the Platform after creating a cohort on Gen3 BioData Catalyst or Gen3 Anvil which includes DRS URIs and a subset of metadata from the PFB file.

In addition, the PFB file that was used as a manifest is also imported into your project since it contains all available data.

The procedure consists of:

  • Selecting and exporting data from BioData Catalyst powered by Gen3 or Gen3 Anvil.
  • Connecting Gen 3 Biodata Catalyst or Gen3 AnVIL account to your account on the Platform. This step occurs the first time you land on the Platform after exporting data from either Gen 3 BDC or Gen3 AnVIL.
  • Importing data into a controlled project on BioData Catalyst powered by Seven Bridges.
  • (Optional) If you want to use raw clinical data in your analyses, unpacking and converting Avro files to JSON.

Selecting and exporting data from BioData Catalyst powered by Gen3

This step takes place on BioData Catalyst powered by Gen3.

Before you can import the data into your project, you first need to export it from BioData Catalyst Gen3 by following the procedure below:

  1. On the top navigation bar click Exploration.
  2. In the pane on the left, select the Data tab.
  3. Use the available filters in the Filters section to select patient data of your interest.
  4. When you're done with filtering, click Export to Seven Bridges and choose option Export to Biodata Catalyst to start the export process.

Selecting and exporting data from Anvil powered by Gen3

This step takes place on Anvil powered by Gen3.

Before you can import the data into your project, you first need to export it from Anvil Gen3 by following the procedure below:

  1. On the top navigation bar click Exploration.
  2. In the pane on the left, select the Data tab.
  3. Use the available filters in the Filters section to select patient data of your interest.
  4. When you're done with filtering, click Export to Seven Bridges and choose option Export to Biodata Catalyst to start the export process.

Importing data into a project on BioData Catalyst powered by Seven Bridges

This step takes place on BioData Catalyst powered by Seven Bridges.

After using the export feature, you will land on a page for setting up the import to the Platform (further discussed below).

  1. Select the project you want to import the data to (Important note: this data can only be imported to a controlled project). Alternatively, you can create a new project and upload the files there.

  1. (Optional) Add tags for the files.
  2. (Optional) Choose the method for resolving naming conflicts in case files with the same names already exist in your project.
  3. Click Import Data.

The files will be imported. The notification in the upper right corner contains information on the exact number of files what were uploaded. You will be able to monitor the progress. This action may take up to 5-10 minutes to complete depending on the size of the file.

To use data found from PFB (avro) in your analyses, the data will first need to be extracted from archives and then converted to JSON, following the procedure below.

📘

Along with the files you have selected, a PFB (avro) file will also be exported which may contain additional information.

Extract and convert Avro files to JSON

This step takes place in Data Studio on BioData Catalyst powered by Seven Bridges.

  1. Access the project containing the data you imported from BioData Catalyst powered by Gen3.
  2. Open the Data Studio tab. This takes you to the Data Studio home page. If you have previous analyses, they will be listed on this page.
  3. In the top-right corner click Create new analysis. The Create new analysis wizard is displayed.
  4. Name your analysis in the Analysis name field.
  5. Select JupyerLab as the analysis environment.
  6. Keep the default Environment setup. Each setup is a preinstalled set of libraries that is tailored for a specific purpose. Learn more.
  7. Keep the default Instance type and Suspend time settings.
  8. Click Start. The Platform will start acquiring an adequate instance for your analysis, which may take a few minutes.
  9. Once the Platform has acquired an instance for your analysis, the JupyterLab home screen is displayed.
  10. In the Notebook section, select Python 3. A new blank Jupyter notebook opens.
  11. Start by installing the fastavro library:
!pip install fastavro
  1. Import the reader feature from fastavro and import the json Python library:
from fastavro import reader
import json
  1. Use the reader functionality of the fastavro library to read the Avro file and assign its content to the file variable. Make sure to replace <avro-file-name> with the name of the Avro file, including '/sbgenomics/project-files/.avro'
file = []
with open('<avro-file-name>.avro', 'rb') as fo:
    avro_reader = reader(fo)
    for record in avro_reader:
        file.append(record)
  1. Save the content to a JSON file. Make sure to replace <json-file-name> with the actual name you want to use for the JSON file.
json_file = []
with open('<json-file-name>.json', 'w') as fp:
    json.dump(file, fp, indent=2)
  1. Finally, save the JSON file to the /sbgenomics/output-files/ directory within the analysis. When you stop the Data Studio analysis, the JSON file will be saved to the project, and will be available for further use in tasks within that project.
!cp <json-file-name>.json /sbgenomics/output-files/