Skip to content

Dask-Jobqueue unable to run on compute node #363

@JatinGharat94

Description

@JatinGharat94

I am trying to run some computation on cray compute nodes using PBSCluster.
I have written the dask-jobqueue script as follows:

import xarray as xr
from dask.distributed import Client
from dask_jobqueue import PBSCluster
from dask import compute
import dask

cluster = PBSCluster(queue='research',
                     project='DaskOnPBS',
                     local_directory='/hpcs_rnd/Python_Data_Analysis/Jatin/dask-worker-space/',
                     cores=1,
                     processes=1,
                     memory='10GB',
                     resource_spec='select=1:ncpus=36:mem=20GB:vntype=cray_compute')

cluster.scale(8)
client = Client(cluster)

But IT administrators told me that my script runs on login node rather than on compute node. They told me to run this script on compute node using aprun command which is similar to mpirun.

Usually job scripts are submitted on our HPC in the following manner:

#!/bin/sh
#PBS -N any_job
#PBS -q research
#PBS -l select=1:ncpus=40:vntype=cray_compute
#PBS -l walltime=500:00:00

cd $WORK_DIRECTORY
aprun -n 40 -N 40 ./your_filename > output.log

I do not know how to use aprun command in PBSCluster to run my script on compute node. Please help!

Metadata

Metadata

Assignees

No one assigned

    Labels

    questionFurther information is requested

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions