wICE quick start guide — VSC documentation (2024)

wICE is the most recent KU Leuven/UHasselt Tier-2 cluster.It can be used for most workloads, and has nodes with a lot of memory, as well asnodes with GPUs.

wICE does not have separate login nodes and can be accessed either from theGenius login nodes, or from your web browser via theOpen On-Demand service.

Running jobs on wICE#

There are several type of nodes in the wICE cluster: normal compute nodes, GPU nodes,big memory nodes and nodes configured for interactive use.The resources specifications for jobs have to be tuned to use these nodes properly.

In general, the maximum walltime for wICE jobs is 3 days (72 hours).Only jobs submitted to the batch_long partition are allowed to have walltimes up to7 days (168 hours), as illustrated below.

Similar to Genius, wICE uses Slurm as the workload manager.More information about converting pbs scripts and commands into Slurm can be foundhere.

A Slurm jobscript for wICE will typically look like this:

#!/bin/bash -l#SBATCH --clusters=wice#SBATCH --partition=...#SBATCH --time=...#SBATCH --nodes=...#SBATCH --ntasks-per-node=...#SBATCH --account=...module use /apps/leuven/icelake/2021a/modules/allmodule load ......

For information about using and installing software on wICE, see theadvanced guide for wICE.

Submit to a compute node#

In these examples, the submission will be done from login nodes of the Genius cluster,therefore you always need to explicitly specify the cluster you want to use.The relevant option for that is -M|--clusters which takes wice or genius asa valid value.

To submit to a compute node you need to provide the required number of nodes and cores.For example to request 2 nodes with each 72 cores for 2 hours you can submit like this:

$ sbatch --account=lp_myproject --clusters=wice --nodes=2 --ntasks-per-node=72 \ --time=2:00:00 myjobscript.slurm

More information about accounting used on wice can be found on the Leuven accountingpage and on KU Leuven credits.

Submit a long job to a compute node#

To submit to a compute node a job longer than 3 days you need to submit specifically to the batch_long partition:

$ sbatch --account=lp_myproject --clusters=wice --nodes=2 --ntasks-per-node=72 \ --time=6-16:00:00 --partition=batch_long myjobscript.slurm

Submit to the interactive partition#

The interactive nodes are located below the interactive partition.The users are allowed to request a maximum of 8 cores, one (shared portion of a) GPU instance,and for a maximum walltime of 16 hours.These nodes are intended for interactive use.Instead of submitting a job script, you open an interactive session on a compute node asfollows:

$ srun --account=lp_myproject --ntasks=1 --time=01:00:00 --partition=interactive \ --clusters=wice --pty bash -l

If a GPU is necessary for the visualization process, it can be requested (max 1 GPU instanceand total of 8 cores for at most 16 hours).The available GPU is a single A100 which has been split in 7 GPU instances (one of whichwill be allocated to your job).Additionally, X11 forwarding should be enabled:

$ srun --account=lp_myproject --nodes=1 --time=16:00:00 --ntasks-per-node=8 \ --gpus-per-node=1 --partition=interactive --clusters=wice --x11 --pty bash -l

Note

The interactive partition is intended for relatively lightweight interactive work,such as compiling software, running small preprocessing scripts, small-scaledebugging, or visualization.This is the reason why the amount of resources you can get in a job is limited on the interactive partition. In case you must do heavy computational work in an interactive way, it is also possible to submit interactive jobs to the other partitions. For instance suppose you need to debug a program using more than 8 cores. In that case you can use the command above to run an interactive job, changing the partition to batch, gpu, or bigmem and adapting the resources as needed. Do note that in general it is recommended to run heavy computational work in a script which you run as a batch job (so without opening an interactive terminal on the compute node).

Note

We urge our users to recompile their software on wICE.Using a software on wICE which is formerly compiled on Genius can lead to confusingerror messages, specifically in case of MPI applications.This is true even for (mini)conda environments.So, as a general rule of thumb, we propose that your software (including conda) berecompiled on wICE and put in a separate folder to distinguish the build from existingsoftware that suit Genius.

Submit to a big memory node#

The big memory nodes (2048GB of RAM) are also located in the bigmem partition.In case of the big memory nodes it is also important to add your memory requirements(the maximum of memory per core that can be requested is 28000MB/core), for example:

$ sbatch --account=lp_myproject --clusters=wice --time=01:00:00--nodes=2 \ --ntasks-per-node=72--partition=bigmem --mem-per-cpu=28000M myjobscript.slurm

Submit to a GPU node#

The GPU nodes are located in the gpu partition, so you will need to explicitly specifyit when submitting your job.Similar to the other nodes, the GPU nodes can be shared by different jobs from different users.However every user will have exclusive access to the number of GPUs requested.If you want to use only 1 GPU of type A100 you can submit for example like this:

$ sbatch --account=lp_myproject --clusters=wice --nodes=1 --ntasks=18 \ --gpus-per-node=1 --partition=gpu myjobscript.slurm

Note that in case of 1 GPU you have to request 18 cores.In case you need more GPUs you have to multiply the 18 cores with the number of GPUsrequested, so in case of for example 3 GPUs you will have to specify this:

$ sbatch --account=lp_myproject --clusters=wice --nodes=1 --ntasks=54 \ --gpus-per-node=3 --partition=gpu myjobscript.slurm
wICE quick start guide — VSC documentation (2024)

References

Top Articles
Latest Posts
Article information

Author: Tyson Zemlak

Last Updated:

Views: 5475

Rating: 4.2 / 5 (43 voted)

Reviews: 90% of readers found this page helpful

Author information

Name: Tyson Zemlak

Birthday: 1992-03-17

Address: Apt. 662 96191 Quigley Dam, Kubview, MA 42013

Phone: +441678032891

Job: Community-Services Orchestrator

Hobby: Coffee roasting, Calligraphy, Metalworking, Fashion, Vehicle restoration, Shopping, Photography

Introduction: My name is Tyson Zemlak, I am a excited, light, sparkling, super, open, fair, magnificent person who loves writing and wants to share my knowledge and understanding with you.