SeaWulf Shared Queues

SeaWulf Shared Queues Guide

Shared queues allow multiple users to run jobs on the same node simultaneously. Proper resource management is crucial to avoid exceeding memory or CPU limits.

Shared Queues Overview

Shared queues are designed for users who do not need exclusive access to a node but still require compute resources. When using shared queues

Available Shared Queues

Queue CPU Architecture Vector/Matrix Extension CPU Cores per Node GPUs per Node Node Memory Default Runtime Max Runtime Max Nodes Min Nodes Max Simultaneous Jobs per User Multiple Users per Node
short-40core-shared Intel Skylake AVX512 40 0 192 GB 1 hour 4 hours 4 n/a n/a Yes
long-40core-shared Intel Skylake AVX512 40 0 192 GB 8 hours 24 hours 3 n/a n/a Yes
extended-40core-shared Intel Skylake AVX512 40 0 192 GB 8 hours 3.5 days 1 n/a n/a Yes
short-96core-shared AMD EPYC Milan AVX2 96 0 256 GB 1 hour 4 hours 4 n/a n/a Yes
long-96core-shared AMD EPYC Milan AVX2 96 0 256 GB 8 hours 24 hours 3 n/a n/a Yes
extended-96core-shared AMD EPYC Milan AVX2 96 0 256 GB 8 hours 3.5 days 1 n/a n/a Yes

Instructions for Using Shared Queues

  • Request memory explicitly: Use #SBATCH --mem=[amount] to avoid exceeding the memory available on a shared node.
  • Check resource limits: Know the total node memory and CPU cores and plan your job accordingly.
  • Monitor usage: Use squeue -u $USER to check your jobs and top or htop to see memory/CPU usage.