Skip to content

Commit 4bbe10d

Browse files
committed
adding documentation info for mi2101x partition
Signed-off-by: Karl W. Schulz <[email protected]>
1 parent 9a2308a commit 4bbe10d

File tree

2 files changed

+10
-7
lines changed

2 files changed

+10
-7
lines changed

docs/hardware.md

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ The HPC Fund Research Cloud consists of 40 high performance computing (HPC) serv
44

55
## Compute servers
66

7-
Each compute server consists of two [AMD EPYC&trade;](https://www.amd.com/en/processors/epyc-server-cpu-family) processors with access to 512 GB (or more) of main memory. High-speed user network connectivity for inter-node communication is accommodated by a [ConnextX-6](https://nvdam.widen.net/s/5j7xtzqfxd/connectx-6-infiniband-datasheet-1987500-r2) MT28908 Infiniband host channel adapter providing a maximum port speed of 200 Gb/s. For accelerated analysis, each node also includes one or more [AMD Instinct&trade;](https://www.amd.com/en/products/accelerators/instinct.html) accelerators. Multiple generations of accelerators are available within the system with key characteristics highlighted as follows:
7+
Each bare-metal compute server consists of two [AMD EPYC&trade;](https://www.amd.com/en/processors/epyc-server-cpu-family) processors with access to 512 GB (or more) of main memory. High-speed user network connectivity for inter-node communication is accommodated by a [ConnextX-6](https://nvdam.widen.net/s/5j7xtzqfxd/connectx-6-infiniband-datasheet-1987500-r2) MT28908 Infiniband host channel adapter providing a maximum port speed of 200 Gb/s. For accelerated analysis, each node also includes one or more [AMD Instinct&trade;](https://www.amd.com/en/products/accelerators/instinct.html) accelerators. Multiple generations of accelerators are available within the system with key characteristics highlighted as follows:
88
<!-- * [AMD MI100 Accelerator](https://www.amd.com/en/products/accelerators/instinct/mi100.html)
99
* Peak double-precision (FP64) performance of 11.5 TFLOPs
1010
* 32 GB of high bandwidth memory (HBM2e)
@@ -22,11 +22,12 @@ Each compute server consists of two [AMD EPYC&trade;](https://www.amd.com/en/pro
2222
* Form factor: OAM Module -->
2323

2424
```{table} Table 1: Hardware Overview of Available Node Types
25-
| Accelerator | Peak FP64 | HBM Capacity | HBM Peak B/W | Host CPU | Host Memory |
26-
| --------- | :------: | :---------: | :---------------: | :------------------------------------------: | :---: |
27-
| [AMD MI210](https://www.amd.com/en/products/accelerators/instinct/mi200/mi210.html) | 45.3 TFLOPs | 64GB | 1.6 TB/s | 2 X EPYC 7V13 64-core | 512 GB |
28-
| [AMD MI250](https://www.amd.com/en/products/accelerators/instinct/mi200/mi250.html) | 45.3 TFLOPs (per GCD) | 64GB (per GCD) | 1.6 TB/s (per GCD) | 2 X EPYC 7763 64-Core | 1.5 TB |
29-
| [AMD MI300X](https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html) | 81.7 TFLOPs | 192GB | 5.3 TB/s | 2 X EPYC 9684X 96-Core | 2.3 TB |
25+
| Accelerator | Peak FP64 | HBM<br> Capacity | HBM <br> Peak B/W | Host CPU | Host<br>Memory |
26+
| --------- | :------: | :---------: | :---------------: | :--------: | :---: |
27+
| [AMD MI210](https://www.amd.com/en/products/accelerators/instinct/mi200/mi210.html) <br> (1&nbsp;GPU/node) | 45.3 TFLOPs | 64GB | 1.6 TB/s | EPYC 7V13 16-core (VM) | 64 GB (VM) |
28+
| [AMD MI210](https://www.amd.com/en/products/accelerators/instinct/mi200/mi210.html) <br> (4&nbsp;GPUs/node) | 45.3 TFLOPs | 64GB | 1.6 TB/s | 2 X EPYC 7V13 64-core | 512 GB |
29+
| [AMD MI250](https://www.amd.com/en/products/accelerators/instinct/mi200/mi250.html) <br> (8&nbsp;GCDs/node) | 45.3 TFLOPs (per GCD) | 64GB (per GCD) | 1.6 TB/s (per GCD) | 2 X EPYC 7763 64-Core | 1.5 TB |
30+
| [AMD MI300X](https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html) <br> (8&nbsp;GPUs/node) | 81.7 TFLOPs | 192GB | 5.3 TB/s | 2 X EPYC 9684X 96-Core | 2.3 TB |
3031
```
3132

3233
Note that one AMD MI250 accelerator provides two Graphics Compute Dies (GCDs) for which the programmer can use as two separate GPUs.

docs/jobs.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ Multiple partitions (or queues) are available for users to choose from and each
1010
| Queue | Max Time | Max Node(s) | Charge Multiplier | Configuration |
1111
| --------- | -------- | ----------- | ----------------- | ------------------------------------------------ |
1212
| `devel` | 30 min. | 1 | 1.0X | Targeting short development needs (4xMI210). |
13+
| `mi2101x` | 12 hours | 1 | 0.25X | 1 MI210 accelerator per node. |
1314
| `mi2104x` | 24 hours | 16 | 1.0X | 4 x MI210 accelerators per node. |
1415
| `mi2508x` | 12 hours | 10 | 1.7X | 4 x MI250 accelerators (8 GPUs) per node. |
1516
| `mi3008x` | 4 hours | 1 | 2.0X | 8 x MI300X accelerators per node. |
@@ -28,6 +29,7 @@ Since multiple generations of Instinct&trade; accelerators are available across
2829
Partition Name | GPU Type | ROCm Offload Architecture Compile Flag
2930
---------------|-----------|-----------------------
3031
devel | MI210 x 4 | `--offload-arch=gfx90a`
32+
mi2101x | MI210 x 1 | `--offload-arch=gfx90a`
3133
mi2104x | MI210 x 4 | `--offload-arch=gfx90a`
3234
mi2508x | MI250 x 8 | `--offload-arch=gfx90a`
3335
mi3008x | MI300 x 8 | `--offload-arch=gfx942`
@@ -270,7 +272,7 @@ a new (or open an existing) notebook and access the GPUs on the compute node:
270272
![jupyter-notebook](images/jupyter-notebook-gpus.PNG)
271273

272274
```{tip}
273-
Please see the [Python Environment](./software.md#python-environment) section to understand how the base Python environment and `pytorch` and `tensorflow` modules can be customized.
275+
Please see the [Python Environment](python-environment) section to understand how the base Python environment and `pytorch` and `tensorflow` modules can be customized.
274276
```
275277

276278
## Large Language Models (Ollama)

0 commit comments

Comments
 (0)