Skip to content

Commit bde043e

Browse files
committed
addition of MI300 availability; fix warnings for two cross-references
Signed-off-by: Karl W. Schulz <[email protected]>
1 parent 6fec89c commit bde043e

File tree

3 files changed

+22
-0
lines changed

3 files changed

+22
-0
lines changed

docs/hardware.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,7 @@ Each compute server consists of two [AMD EPYC&trade;](https://www.amd.com/en/pro
2727
| [AMD MI100](https://www.amd.com/en/products/accelerators/instinct/mi100.html) | 11.5 TFLOPs | 32GB | 1.2 TB/s | 2 X EPYC 7V13 64-core | 512 GB |
2828
| [AMD MI210](https://www.amd.com/en/products/accelerators/instinct/mi200/mi210.html) | 45.3 TFLOPs | 64GB | 1.6 TB/s | 2 X EPYC 7V13 64-core | 512 GB |
2929
| [AMD MI250](https://www.amd.com/en/products/accelerators/instinct/mi200/mi250.html) | 45.3 TFLOPs (per GCD) | 64GB (per GCD) | 1.6 TB/s (per GCD) | 2 X EPYC 7763 64-Core | 1.5 TB |
30+
| [AMD MI300X](https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html) | 81.7 TFLOPs | 192GB | 5.3 TB/s (per GCD) | 2 X EPYC 9684X 96-Core | 2.3 TB |
3031
```
3132

3233
Note that one AMD MI250 accelerator provides two Graphics Compute Dies (GCDs) for which the programmer can use as two separate GPUs.

docs/jobs.md

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,10 +13,29 @@ Multiple partitions (or queues) are available for users to choose from and each
1313
| `mi1008x` | 24 hours | 5 | 0.8X | 8 x MI100 accelerators per node. |
1414
| `mi2104x` | 24 hours | 16 | 1.0X | 4 x MI210 accelerators per node. |
1515
| `mi2508x` | 12 hours | 10 | 1.7X | 4 x MI250 accelerators (8 GPUs) per node. |
16+
| `mi3008x` | 4 hours | 1 | 2.0X | 8 x MI300X accelerators per node. |
17+
| `mi3008x_long` | 8 hours | 1 | 2.0X | 8 x MI300X accelerators per node. |
1618
```
1719

1820
Note that special requests that extend beyond the above queue limits may potentially be accommodated on a case-by-case basis. You must have an active accounting allocation in order to submit jobs and the resource manager will track the combined number of **node** hours consumed by each job and deduct the [total node hours]*[charge multiplier] from your available balance.
1921

22+
23+
## Offload Architecture Options
24+
25+
Since multiple generations of Instinct&trade; accelerators are available across the cluster, users building their own [HIP](https://rocm.docs.amd.com/projects/HIP/en/latest/) applications should include the correct target offload architecture during compilation based on the desired GPU type. The following table highlights the offload architecture types and compilation option that maps to available SLURM partitions.
26+
27+
```{table} Table 2: Offload architecture settings for local HIP compilation
28+
:widths: 25 25 50
29+
Partition Name | GPU Type | ROCm Offload Architecture Compile Flag
30+
---------------|-----------|-----------------------
31+
devel | MI210 x 4 | `--offload-arch=gfx90a`
32+
mi2104x | MI210 x 4 | `--offload-arch=gfx90a`
33+
mi2508x | MI250 x 8 | `--offload-arch=gfx90a`
34+
mi3008x | MI300 x 8 | `--offload-arch=gfx942`
35+
mi3008x_long | MI300 x 8 | `--offload-arch=gfx942`
36+
mi1008x | MI100 x 8 | `--offload-arch=gfx908`
37+
```
38+
2039
## Batch job submission
2140

2241
Example SLURM batch job submission scripts are available on the login node at `/opt/ohpc/pub/examples/slurm`. A basic starting job for MPI-based applications is available in this directory named `job.mpi` and is shown below for reference:
@@ -162,6 +181,7 @@ The table below highlights several of the more common user-facing SLURM commands
162181
| scontrol | view or modify a job configuration |
163182
```
164183

184+
(jupyter)=
165185
## Jupyter
166186

167187
Users can run Jupyter Notebooks on the HPC Fund compute nodes by making a copy

docs/software.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,7 @@ The Lmod system provides a flexible mechanism to manage your local software envi
3030
The `module help` command can also be run locally on the system to get more information on available Lmod options and sub-commands.
3131
```
3232

33+
(python-environment)=
3334
## Python Environment
3435

3536
A base Python installation is available on the HPC Fund cluster which includes a handful of common packages (e.g., `numpy`, `pandas`). If additional packages are needed, users can customize their environments by installing packages with a user install, creating a Python virtual environment to install packages in, or loading a module for a specific package (e.g., `pytorch`, `tensorflow`). Examples of each method are given below.

0 commit comments

Comments
 (0)