Skip to content

Commit

Permalink
[manual] complete my proofreading for 7.0.
Browse files Browse the repository at this point in the history
Add access links to Digital Alliance.
Clarify -n vs -c and their --ntasks and --cpus-per-tasks a bit.
Fix some typos.
  • Loading branch information
smokhov committed Nov 12, 2023
1 parent c1f9deb commit 401ccc6
Show file tree
Hide file tree
Showing 4 changed files with 36 additions and 14 deletions.
4 changes: 2 additions & 2 deletions doc/scheduler-directives.tex
Original file line number Diff line number Diff line change
Expand Up @@ -43,7 +43,6 @@ \subsubsection{Directives}
#SBATCH [email protected]
#SBATCH --chdir=./ ## Use current directory as working directory
#SBATCH -N 1
#SBATCH --ntasks=1
#SBATCH -n 8 ## Request 8 cores
#SBATCH --mem=32G ## Assign 32G memory per node
\end{verbatim}
Expand All @@ -55,10 +54,11 @@ \subsubsection{Directives}
(multithreaded) parallelism (i.e., cores) you attach to your job (up to 32 by default).
%be sure to delete or comment out the \verb| #$ -pe smp | parameter if it
%is not relevant;

Replace, \verb+<memory>+, with the value (in GB), that you want
your job's memory space to be (up to 500 depending on the node), and all jobs MUST have a memory-space
assignment.

%
If you are unsure about memory footprints, err on assigning a generous
memory space to your job, so that it does not get prematurely terminated.
%(the value given to \api{h\_vmem} is a hard memory ceiling).
Expand Down
10 changes: 5 additions & 5 deletions doc/scheduler-job-examples.tex
Original file line number Diff line number Diff line change
Expand Up @@ -105,14 +105,14 @@ \subsection{Scheduling On The GPU Nodes}
%#$ -l gpu=[1|2]
%\end{verbatim}
\begin{verbatim}
#BATCH --gpu=[1|2]
#SBATCH --gpus=[1|2]
\end{verbatim}

% TODO: verify accuracy
Single-GPU jobs are granted 5~CPU cores and 80GB of system memory, and
dual-GPU jobs are granted 10~CPU cores and 160GB of system memory. A
total of \emph{four} GPUs can be actively attached to any one user at any given
time.
% Single-GPU jobs are granted 5~CPU cores and 80GB of system memory, and
% dual-GPU jobs are granted 10~CPU cores and 160GB of system memory. A
% total of \emph{four} GPUs can be actively attached to any one user at any given
% time.

Once that your job script is ready, you can submit it to the GPU partition (queue)
with:
Expand Down
10 changes: 10 additions & 0 deletions doc/scheduler-scripting.tex
Original file line number Diff line number Diff line change
Expand Up @@ -409,6 +409,16 @@ \subsection{Requesting Multiple Cores (i.e., Multithreading Jobs)}
core count available, please (please) request 32 cores, to avoid node
oversubscription (i.e., to avoid overloading the CPUs).
\textbf{Important} note about \option{--ntasks} or \option{--ntasks-per-node}
(\option{-n}) talks about processes (usually the ones ran with \tool{srun}).
\option{--cpus-per-task} (\option{-c}) corresponds to threads per process.
Some programs consider them equivalent, some don't. Fluent for example
uses \option{--ntasks-per-node=8} and \option{--cpus-per-task=1},
some just set \option{--cpus-per-task=8} and \option{--ntasks-per-node=1}.
If one of them is not $1$ then some applications need to be told to
use $n*c$ total cores.
Core count associated with a job appears under,
%``states'', in the, \texttt{qstat -f -u "*"}, output.
``AllocCPUS'', in the, \texttt{qacct -j}, output.
Expand Down
26 changes: 19 additions & 7 deletions doc/speed-manual.tex
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,8 @@
%\date{Version 6.5}
%\date{\textbf{Version 6.6-dev-07}}
%\date{\textbf{Version 6.6} (final GE version)}
\date{\textbf{Version 7.0-dev-01}}
%\date{\textbf{Version 7.0-dev-01}}
\date{\textbf{Version 7.0}}

% Authors are joined by \and and their affiliations are on the
% subsequent lines separated by \\ just like the article class
Expand Down Expand Up @@ -129,7 +130,7 @@ \subsection{Resources}

% TODO: for now comment out for 7.0; if when we update that
% preso, we will re-link it here. However, keep the citation.
\nocite{speed-intro-preso}.
\nocite{speed-intro-preso}
%\item
%\href
% {https://docs.google.com/presentation/d/1zu4OQBU7mbj0e34Wr3ILXLPWomkhBgqGZ8j8xYrLf44}
Expand All @@ -155,7 +156,9 @@ \subsection{Team}
\end{itemize}

\noindent
We receive support from the rest of AITS teams, such as NAG, SAG, FIS, and DOG.
We receive support from the rest of AITS teams, such as NAG, SAG, FIS, and DOG.\\
%
\url{https://www.concordia.ca/ginacody/aits.html}

% ------------------------------------------------------------------------------
\subsection{What Speed Consists of}
Expand All @@ -175,7 +178,7 @@ \subsection{What Speed Consists of}
256GB of RAM.

\item
7 new SPEED2 servers with 64 CPU cores each 4x A100 80GB GPUs, partitioned
7 new SPEED2 servers with 64 CPU cores each 4x A100 80~GB GPUs, partitioned
into 4x 20GB each; larger local storage for TMPDIR.

\item
Expand All @@ -202,7 +205,10 @@ \subsection{What Speed Is Ideal For}

\begin{itemize}
\item
To design and develop, test and run parallel, batch, and other algorithms, scripts with partial data sets. ``Speed'' has been optimised for compute jobs that are multi-core aware, require a large memory space, or are iteration intensive.
To design and develop, test and run parallel, batch, and other algorithms,
scripts with partial data sets. ``Speed'' has been optimised for compute jobs
that are multi-core aware, require a large memory space, or are iteration
intensive.
\item
Prepare them for big clusters:
\begin{itemize}
Expand All @@ -222,7 +228,7 @@ \subsection{What Speed Is Ideal For}
\item
CPU-based jobs.
\item
CUDA GPU jobs (\texttt{speed-05}, \texttt{speed-17}, \texttt{speed-37}--\texttt{speed-43}).
CUDA GPU jobs (\texttt{speed-01|-03|-05}, \texttt{speed-17}, \texttt{speed-37}--\texttt{speed-43}).
\item
Non-CUDA GPU jobs using OpenCL (\texttt{speed-19} and \texttt{-01|03|05|17|25|27|37-43}).
\end{itemize}
Expand Down Expand Up @@ -892,7 +898,13 @@ \section{Sister Facilities}
Contact Thomas Beaudry for details and how to obtain access.
\item
Digital Research Alliance Canada (Compute Canada / Calcul Quebec),\\
\url{https://alliancecan.ca/}
\url{https://alliancecan.ca/}. Follow
\href
{https://alliancecan.ca/en/services/advanced-research-computing/account-management/apply-account}
{this link}
on the information how to obtain access (students need to be sponsored
by their supervising faculty members, who should create accounts
first). Their SLURM examples are here: \url{https://docs.alliancecan.ca/wiki/Running_jobs}

\end{itemize}

Expand Down

0 comments on commit 401ccc6

Please sign in to comment.