User Tools

Site Tools


slurm

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revisionPrevious revision
Next revision
Previous revision
slurm [2021/01/15 06:07] – Add array jobs explanation. kameranisslurm [2022/10/07 15:13] (current) borja
Line 8: Line 8:
 Slurm is similar to most other queue systems in that you write a batch script, then submit it to the queue manager. The queue manager schedules your job to run on the queue (or partition in Slurm parlance) that you designate. Below is an outline of how to submit jobs to Slurm, how Slurm decides when to schedule your job, and how to monitor progress. Slurm is similar to most other queue systems in that you write a batch script, then submit it to the queue manager. The queue manager schedules your job to run on the queue (or partition in Slurm parlance) that you designate. Below is an outline of how to submit jobs to Slurm, how Slurm decides when to schedule your job, and how to monitor progress.
  
 +
 +===== Communication =====
 +==== Mailing List ====
 +If you are going to be a user of any of the CS Slurm clusters please sign up for the mailing list. Downtime and other relevant information will be announced here.
 +
 +[[ https://mailman.cs.uchicago.edu/cgi-bin/mailman/listinfo/slurm | Mailing List ]]
 +
 +==== Discord ====
 +There is a dedicated text channel ''%%#slurm%%'' on the UChicago CS Discord server. Please note that this Discord server is //only// for UChicago-affiliated users. You can find a link to our Discord server on the [[start|front page]] of this wiki.
  
 ===== Clusters ===== ===== Clusters =====
 +
 +We have a couple different clusters. If you don't know where to start please use the ''%%Peanut%%'' cluster. The ''%%AI Cluster%%'' is for GPU jobs and more advanced users.
  
   * [[slurm:peanut|Peanut Cluster]]   * [[slurm:peanut|Peanut Cluster]]
Line 24: Line 35:
  
 To use this cluster there are specific nodes you need to log into. Please visit the dedicated AI cluster page for more information. To use this cluster there are specific nodes you need to log into. Please visit the dedicated AI cluster page for more information.
- 
- 
- 
  
  
Line 35: Line 43:
   ssh user@linux.cs.uchicago.edu   ssh user@linux.cs.uchicago.edu
  
-If you want to use the AI Cluster you will need to login into:+If you want to use the AI Cluster you will need to have previously requested access by sending in a ticket. Afterwards, you may login into:
  
   ssh user@fe.ai.cs.uchicago.edu   ssh user@fe.ai.cs.uchicago.edu
Line 41: Line 49:
 Please read up on the specifics on the cluster you are interested in. Please read up on the specifics on the cluster you are interested in.
  
-===== Mailing List ===== 
-If you are going to be a user of this cluster please sign up for the mailing list. Downtime and other relevant information will be announced here. 
  
-[[ https://mailman.cs.uchicago.edu/cgi-bin/mailman/listinfo/slurm | Mailing List ]] 
  
 ===== Documentation ===== ===== Documentation =====
Line 342: Line 347:
 Please make sure you specify $CUDA_HOME and if you want to take advantage of CUDNN libraries you will need to append /usr/local/cuda-x.x/lib64 to the $LD_LIBRARY_PATH environment variable. Please make sure you specify $CUDA_HOME and if you want to take advantage of CUDNN libraries you will need to append /usr/local/cuda-x.x/lib64 to the $LD_LIBRARY_PATH environment variable.
  
-  cuda_version=9.2+  cuda_version=11.1
   export CUDA_HOME=/usr/local/cuda-${cuda_version}   export CUDA_HOME=/usr/local/cuda-${cuda_version}
   export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$CUDA_HOME/lib64   export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$CUDA_HOME/lib64
Line 358: Line 363:
 The variable name is actually misleading; since it does NOT mean the amount of devices, but rather the physical device number assigned by the kernel (e.g. /dev/nvidia2). The variable name is actually misleading; since it does NOT mean the amount of devices, but rather the physical device number assigned by the kernel (e.g. /dev/nvidia2).
  
-For example: If you requested multiple gpu's from Slurm (--gres=gpu:2), the CUDA_VISIBLE_DEVICES variable should contain two numbers(0-3 in this case) separated by a comma (e.g. 1,3).+For example: If you requested multiple gpu's from Slurm (--gres=gpu:2), the CUDA_VISIBLE_DEVICES variable should contain two numbers(0-3 in this case) separated by a comma (e.g. 0,1). 
 + 
 +The numbering is relative and specific to you. For example: two users with one job which require two gpus each could be assigned non-sequential gpu numbers. However CUDA_VISIBLE_DEVICES will look like this for both users: 0,1 
  
  
/var/lib/dokuwiki/data/attic/slurm.1610712448.txt.gz · Last modified: 2021/01/15 06:07 by kameranis

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki