Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .github/workflows/docker.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ on:
permissions: {}

env:
VERSION: 0.1.6
VERSION: 0.1.7
IMAGE_NAME: cpg-flow-seqr-loader
DOCKER_DEV: australia-southeast1-docker.pkg.dev/cpg-common/images-dev
DOCKER_MAIN: australia-southeast1-docker.pkg.dev/cpg-common/images
Expand Down
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ CPG-Flow workflows are operated entirely by defining input Cohorts (see [here](h
```bash
analysis-runner \
--skip-repo-checkout \
--image australia-southeast1-docker.pkg.dev/cpg-common/images/cpg-flow-seqr-loader:0.1.6 \
--image australia-southeast1-docker.pkg.dev/cpg-common/images/cpg-flow-seqr-loader:0.1.7 \
--config src/cpg_seqr_loader/config_template.toml \
--config cohorts.toml \ # containing the inputs_cohorts and sequencing_type
--dataset seqr \
Expand All @@ -70,7 +70,7 @@ analysis-runner \
```bash
analysis-runner \
--skip-repo-checkout \
--image australia-southeast1-docker.pkg.dev/cpg-common/images/cpg-flow-seqr-loader:0.1.6 \
--image australia-southeast1-docker.pkg.dev/cpg-common/images/cpg-flow-seqr-loader:0.1.7 \
--config src/cpg_seqr_loader/config_template.toml \
--config cohorts.toml \ # containing the inputs_cohorts and sequencing_type
--dataset seqr \
Expand Down
4 changes: 2 additions & 2 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ description='Seqr-Loader (gVCF-combiner) implemented in CPG-Flow'
readme = "README.md"
# currently cpg-flow is pinned to this version
requires-python = ">=3.10,<3.11"
version="0.1.6"
version="0.1.7"
license={"file" = "LICENSE"}
classifiers=[
'Environment :: Console',
Expand Down Expand Up @@ -122,7 +122,7 @@ hail = ["hail"]
"src/cpg_seqr_loader/scripts/annotate_cohort.py" = ["E501"]

[tool.bumpversion]
current_version = "0.1.6"
current_version = "0.1.7"
parse = "(?P<major>\\d+)\\.(?P<minor>\\d+)\\.(?P<patch>\\d+)"
serialize = ["{major}.{minor}.{patch}"]
commit = true
Expand Down
6 changes: 4 additions & 2 deletions src/cpg_seqr_loader/jobs/CombineGvcfsIntoVds.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
import loguru
from cpg_flow import targets
from cpg_flow import utils as cpg_flow_utils
from cpg_utils import Path, config, hail_batch
from cpg_utils import Path, config, hail_batch, to_path

from cpg_seqr_loader import utils

Expand All @@ -15,13 +15,15 @@ def create_combiner_jobs(
multicohort: targets.MultiCohort,
output_vds: Path,
combiner_plan: Path,
temp_dir: Path,
temp_dir_string: str,
job_attrs: dict[str, str],
) -> 'BashJob | None':
vds_path: str | None = None
sg_ids_in_vds: set[str] = set()
sgs_to_remove: list[str] = []

temp_dir = to_path(temp_dir_string)

# check for a VDS by ID - this is not the typical RD process
if vds_id := config.config_retrieve(['workflow', 'use_specific_vds'], None):
vds_result_or_none = utils.query_for_specific_vds(vds_id)
Expand Down
4 changes: 2 additions & 2 deletions src/cpg_seqr_loader/stages.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ class CombineGvcfsIntoVds(stage.MultiCohortStage):
def expected_outputs(self, multicohort: targets.MultiCohort) -> dict[str, Path | str]:
return {
'vds': self.prefix / f'{multicohort.name}.vds',
'tmp': self.tmp_prefix / 'temp_dir',
'tmp': str(self.tmp_prefix / 'temp_dir'),
}

def queue_jobs(self, multicohort: targets.MultiCohort, inputs: stage.StageInput) -> stage.StageOutput:
Expand All @@ -45,7 +45,7 @@ def queue_jobs(self, multicohort: targets.MultiCohort, inputs: stage.StageInput)
multicohort=multicohort,
output_vds=outputs['vds'],
combiner_plan=self.tmp_prefix / 'combiner_plan.json',
temp_dir=outputs['tmp'],
temp_dir_string=outputs['tmp'],
job_attrs=self.get_job_attrs(multicohort),
)
return self.make_outputs(multicohort, data=outputs, jobs=job)
Expand Down