-
-
Notifications
You must be signed in to change notification settings - Fork 149
Closed
Labels
questionFurther information is requestedFurther information is requested
Description
I am trying to run some computation on cray compute nodes using PBSCluster.
I have written the dask-jobqueue script as follows:
import xarray as xr
from dask.distributed import Client
from dask_jobqueue import PBSCluster
from dask import compute
import dask
cluster = PBSCluster(queue='research',
project='DaskOnPBS',
local_directory='/hpcs_rnd/Python_Data_Analysis/Jatin/dask-worker-space/',
cores=1,
processes=1,
memory='10GB',
resource_spec='select=1:ncpus=36:mem=20GB:vntype=cray_compute')
cluster.scale(8)
client = Client(cluster)But IT administrators told me that my script runs on login node rather than on compute node. They told me to run this script on compute node using aprun command which is similar to mpirun.
Usually job scripts are submitted on our HPC in the following manner:
#!/bin/sh
#PBS -N any_job
#PBS -q research
#PBS -l select=1:ncpus=40:vntype=cray_compute
#PBS -l walltime=500:00:00
cd $WORK_DIRECTORY
aprun -n 40 -N 40 ./your_filename > output.logI do not know how to use aprun command in PBSCluster to run my script on compute node. Please help!
Metadata
Metadata
Assignees
Labels
questionFurther information is requestedFurther information is requested