Cryo-ET Data Import Guide
SABER leverages copick to provide a flexible and unified interface for accessing tomographic data, whether it's stored locally or remotely on HPC servers or on our CryoET Data Portal. This guide explains how to work with both data sources.
Getting Started
If you need help creating configuration files, detailed tutorials are available:
- Copick Quickstart - Basic configuration and setup
- Data Portal Tutorial - Working with CryoET Data Portal
Starting a New Copick Project
The copick configuration file points to a directory that stores all tomograms, coordinates, and segmentations in an overlay root. Generate a config file using the command line:
copick config filesystem --overlay-root /path/to/overlay
For cryo-ET workflows involving protein 3D coordinate annotation, you can define biological objects during project creation:
copick config filesystem \
--overlay-root /path/to/overlay \
--objects ribosome,True,130,6QZP \
--objects apoferritin,True,65 \
--objects membrane,False
This structure supports both particle picking for sub-tomogram averaging and broader 3D segmentation tasks. Our deep learning platform Octopi π is designed to train models from copick projects for:
- Object 3D localization and particle picking
- Volumetric segmentation of cellular structures
- General 3D dataset annotation and analysis
Directory Structure:
- Overlay root: Writable directory where new results can be added, modified, or deleted
- Static root: Read-only directory that never gets manipulated (frozen data)
π‘ Example Copick Config File (config.json)
The resulting config.json
file would look like this:
{
"name": "test",
"description": "A test project description.",
"version": "1.0.0",
"pickable_objects": [
{
"name": "ribosome",
"is_particle": true,
"label": 1,
"radius": 130,
"pdb_id": "6QZP"
},
{
"name": "apoferritin",
"is_particle": true,
"label": 2,
"radius": 65
},
{
"name": "membrane",
"is_particle": false,
"label": 3
}
],
"overlay_root": "local:///path/to/overlay",
"overlay_fs_args": {
"auto_mkdir": true
},
"static_root": "local:///path/to/static",
"static_fs_args": {
"auto_mkdir": true
}
}
Path Types:
- Local paths:
local:///path/to/directory
- Remote paths:
ssh://server/path/to/directory
The copick config filesystem
command assumes local paths, but you can edit the config file to specify remote locations.
Starting a Copick Project Linked to the Data Portal
Create a copick project that automatically syncs with the CryoET Data Portal:
copick config dataportal --dataset-id DATASET_ID --overlay-root /path/to/overlay
This command generates a config file that syncs data from the portal with local or remote repositories. You only need to specify the dataset ID and the overlay or static path - pickable objects will automatically be populated from the dataset.
Benefits:
- Automatically populates pickable objects from the dataset
- Seamless integration with portal data
- Combines remote portal data with local overlay storage
Importing Local MRC Files
Prerequisites
This workflow assumes:
- All tomogram files are in a flat directory structure (single folder)
- Files are in MRC format (
*.mrc
)
Import Command
If you have tomograms stored locally in *.mrc
format (e.g., from Warp, IMOD, or AreTomo), you can import them into a copick project:
copick add tomogram \
--config config.json \
--tomo-type denoised \
--voxel-size 10 \
--no-create-pyramid \
'path/to/volumes/*.mrc'
Import Parameters Explained
--config config.json
: Path to your copick configuration file--tomo-type denoised
: Specifies the tomogram type (options:raw
,denoised
,filtered
)--voxel-size 10
: Sets voxel size in Γ ngstrΓΆms (10 Γ = 1 nm recommended)--no-create-pyramid
: Skips pyramid generation for faster import'path/to/volumes/*.mrc'
: Path to your MRC file(s) - supports wildcards
Advanced Import Options
If your data doesn't meet the standard requirements (flat directory structure + MRC format), please refer to our Advanced Import Workflows documentation, which covers:
- Nested directory structures
- Different file formats (TIFF, HDF5, etc.)
- Custom import scripts
- Batch processing workflows