You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+4-21Lines changed: 4 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,6 @@
3
3
CloudAI benchmark framework aims to develop an industry standard benchmark focused on grading Data Center (DC) scale AI systems in the Cloud. The primary motivation is to provide automated benchmarking on various systems.
4
4
5
5
## Get Started
6
-
**Note**: instructions for setting up access for `enroot` are available [here](#set-up-access-to-the-private-ngc-registry).
7
6
8
7
Using `uv` tool allows users to run CloudAI without manually managing required Python versions and dependencies.
9
8
```bash
@@ -12,9 +11,12 @@ cd cloudai
12
11
uv run cloudai --help
13
12
```
14
13
14
+
Please refer to the [installation guide](https://nvidia.github.io/cloudai/workloads_requirements_installation.html) for details on setting up workloads' requirements.
15
+
15
16
For details and `pip`-based installation, please refer to the [documentation](https://nvidia.github.io/cloudai/#get-started).
16
17
17
18
## Key Concepts
19
+
18
20
CloudAI operates on four main schemas:
19
21
20
22
-**System Schema**: Describes the system, including the scheduler type, node list, and global environment variables.
@@ -41,25 +43,6 @@ These schemas enable CloudAI to be flexible and compatible with different system
41
43
42
44
For more detailed information, please refer to the [official documentation](https://nvidia.github.io/cloudai/workloads/index.html).
43
45
44
-
## Details
45
-
### Set Up Access to the Private NGC Registry
46
-
First, ensure you have access to the Docker repository. Follow the following steps:
47
-
48
-
1.**Sign In**: Go to [NVIDIA NGC](https://ngc.nvidia.com/signin) and sign in with your credentials.
49
-
2.**Generate API Key**:
50
-
- On the top right corner, click on the dropdown menu next to your profile
51
-
- Select "Setup"
52
-
- In the "Setup" section, find "Keys/Secrets"
53
-
- Click "Generate API Key" and confirm when prompted. A new API key will be presented
54
-
-**Note**: Save this API key locally as you will not be able to view it again on NGC
55
-
56
-
Next, set up your enroot credentials. Ensure you have the correct credentials under `~/.config/enroot/.credentials`:
Replace `<api-key>` with your respective credentials. Keep `$oauthtoken` as is.
61
-
62
-
63
46
## CloudAI Modes Usage Examples
64
47
65
48
CloudAI supports five modes:
@@ -71,7 +54,7 @@ CloudAI supports five modes:
71
54
72
55
### install
73
56
74
-
To install test prerequisites, run CloudAI CLI in install mode.
57
+
To install test prerequisites, run CloudAI CLI in install mode. For more details, please refer to the [installation guide](https://nvidia.github.io/cloudai/workloads_requirements_installation.html).
75
58
76
59
Please make sure to use the correct system configuration file that corresponds to your current setup for installation and experiments.
Copy file name to clipboardExpand all lines: doc/index.md
+7-20Lines changed: 7 additions & 20 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,14 +3,14 @@
3
3
CloudAI benchmark framework aims to develop an industry standard benchmark focused on grading Data Center (DC) scale AI systems in the Cloud. The primary motivation is to provide automated benchmarking on various systems.
4
4
5
5
## Get Started
6
-
**Note**: instructions for setting up access for `enroot` are available [here](set-up-access-to-the-private-ngc-registry).
7
-
8
6
```bash
9
7
git clone git@github.com:NVIDIA/cloudai.git
10
8
cd cloudai
11
9
uv run cloudai --help
12
10
```
13
11
12
+
**Note**: instructions for setting up access for `enroot` are available [installation guide](./workloads_requirements_installation.rst).
13
+
14
14
### `pip`-based installation
15
15
See required Python version in the `.python-version` file, please ensure you have it installed (see how a custom python version [can be installed](#install-custom-python-version)). Follow these steps:
16
16
```bash
@@ -58,24 +58,7 @@ These schemas enable CloudAI to be flexible and compatible with different system
58
58
|SlurmContainer|✅|❌|❌|❌|
59
59
|MegatronRun (experimental)|✅|❌|❌|❌|
60
60
61
-
## Details
62
-
(set-up-access-to-the-private-ngc-registry)=
63
-
### Set Up Access to the Private NGC Registry
64
-
First, ensure you have access to the Docker repository. Follow the following steps:
65
61
66
-
1.**Sign In**: Go to [NVIDIA NGC](https://ngc.nvidia.com/signin) and sign in with your credentials.
67
-
2.**Generate API Key**:
68
-
- On the top right corner, click on the dropdown menu next to your profile
69
-
- Select "Setup"
70
-
- In the "Setup" section, find "Keys/Secrets"
71
-
- Click "Generate API Key" and confirm when prompted. A new API key will be presented
72
-
-**Note**: Save this API key locally as you will not be able to view it again on NGC
73
-
74
-
Next, set up your enroot credentials. Ensure you have the correct credentials under `~/.config/enroot/.credentials`:
Replace `<api-key>` with your respective credentials. Keep `$oauthtoken` as is.
79
62
80
63
## CloudAI Modes Usage Examples
81
64
@@ -89,7 +72,7 @@ CloudAI supports five modes:
89
72
(install)=
90
73
### install
91
74
92
-
To install test prerequisites, run CloudAI CLI in install mode.
75
+
To install test prerequisites, run CloudAI CLI in install mode. For more details, please refer to the [installation guide](./workloads_requirements_installation.rst).
93
76
94
77
Please make sure to use the correct system configuration file that corresponds to your current setup for installation and experiments.
Replace `<api-key>` with your respective credentials. Keep `$oauthtoken` as is.
29
+
30
+
31
+
🤗 Hugging Face models
32
+
----------------------
33
+
34
+
Some workloads require Hugging Face models. CloudAI will download the models from Hugging Face and cache them in the location specified by System's ``hf_home_path`` field. By default, it is set to ``<INSTALL_DIR>/huggingface``, but any other location can be specified. When Slurm is used, this location will be mounted to the container.
35
+
36
+
37
+
Authentication with Hugging Face
38
+
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
39
+
40
+
As of now, CloudAI doesn't handle authentication with Hugging Face, so it is up to the user to enable authentication with Hugging Face in the shell where CloudAI is run. One might need to run the following command:
41
+
42
+
.. code-block:: bash
43
+
44
+
uv run hf auth login
45
+
46
+
Once done, all Hugging Face models will be downloaded using existing authentication.
0 commit comments