Meta-disco is a project focused on metadata inference for biological data files, leveraging both code and AI technologies (LLMs) to automatically extract and validate experimental metadata for the AnVIL Explorer and Terra Data Repository. The goal is to enhance the discoverability and usability of genomic and transcriptomic datasets by ensuring they have accurate, structured metadata.
The meta-disco project consists of two main components:
- LLM Component: Responsible for running the Large Language Models to extract metadata from biological data files.
- Schema Validation Component: Responsible for validating the extracted metadata against LinkML schemas.
The LLM component uses Conda for environment management and Ollama for running the LLM models. Setup instructions can be found in the "Terra Jupyter Ollama Setup" section below.
The Schema Validation component is now located in the schema/
directory and uses Poetry for dependency management. For setup and usage instructions, see the README in the schema/
directory.
A core component of meta-disco is its schema-based approach to metadata validation. The project uses LinkML (Linked Data Modeling Language) to define schemas that specify the expected structure and constraints for metadata associated with biological data files.
The schema defines the structure and constraints for metadata, including:
- Reference Assembly: Specifies the genome reference assembly used (GRCh37, GRCh38, CHM13)
- Data Modality: Indicates the type of biological data (genomic, transcriptomic)
- File Identifiers: Unique identifiers for files in the repository
- Filenames: Names of the data files
These schemas serve two critical purposes:
- They provide a structured format that can be used in prompt creation for LLMs/AI models
- They enable syntactic validation of the metadata predictions
Meta-disco uses the LinkML validation framework to perform syntactic validation of metadata. This ensures that the metadata inferred by AI models or manually entered adheres to the defined schema constraints.
The validator checks:
- Required fields are present
- Values conform to specified data types
- Enumerated values (like reference assemblies) are from the allowed set
- Relationships between metadata elements are consistent
This project has two separate setup processes:
# Clone the repository
git clone https://github.com/DataBiosphere/meta-disco.git
cd meta-disco
# NOTE: Assumes user is in Jupyter Lab container (section: Terra Jupyter Ollama Setup)
# Set up the LLM component package managed through conda and pip
./setup_conda.sh
# Activate metadisco environmen
./post_setup.sh
# Set up the Schema Validation component
cd schema
./setup.sh
The validate
command checks if metadata files conform to the schema:
# Navigate to the schema directory
cd schema
# Using Poetry
poetry run python scripts/validate_outputs.py path/to/metadata.yaml
This validation is crucial for ensuring that metadata inferred by AI models is syntactically correct before it's incorporated into the AnVIL Explorer or Terra Data Repository.
schema/
: Contains the LinkML schema validation componentsrc/meta_disco/schema/
: LinkML schema definitionsscripts/
: Validation scripts
src/meta_disco/
: Main project source codescripts/
: Utility scripts for metadata inference
This section provides instructions to set up and run the terra-jupyter-ollama Docker container on an interactive GPU node managed by the SLURM workload manager.
1. Start an Interactive Node
With GPU support (e.g., with NVIDIA A100 GPUs)
# interactive session with gpu support
$ srun --ntasks=1 \
--cpus-per-task=16 \
--mem=32G \
--gres=gpu:1 \
--partition=gpu \
--nodelist=phoenix-00 \ # specify a gpu node
--time=02:00:00 \
--pty bash
or with only CPU:
# interactive session with no gpu
$ srun --ntasks=1 \
--cpus-per-task=16 \
--mem=32G \
--time=02:00:00 \
--partition=medium \
--pty bash
2. Build the Docker Container
Once on the interactive node, build the Docker image:
docker build -t terra-jupyter-ollama .
3. Run the Docker Container
After building the image, run the container with GPU access, mounted volumes, and port forwarding:
docker run -d --rm --user $(id -u):$(id -g) \
--cpus "${SLURM_CPUS_PER_TASK}" \
--memory 32G \
--gpus="\"device=${SLURM_JOB_GPUS}\"" \
-e HOME=/userhome \
-v "${STORAGE_LOCATION}:/userhome/.ollama" \
-p "${EXPOSED_PORT}:11434" \--name "${CONTAINER_NAME}" ollama/ollama
4. SSH Tunnel to Phoenix
To access the JupyterLab and Ollama services from your local machine, set up an SSH tunnel:
$ ssh -N -L 8889:localhost:8889 \
-L 11434:localhost:11434 \
-J genomics-institute@mustard.prism genomics-institute@phoenix-00
Once connected, you can open:
http://localhost:8889/notebooks/lab for JupyterLab
http://localhost:11434 for Ollama
5. Configure Environment
Open a terminal in JupyterLab and cd to the work dir to create the conda environment.
# Set up the LLM component package managed through conda and pip
./setup_conda.sh
# Activate metadisco environmen
./post_setup.sh