Skip to content

Commit ee1d7c3

Browse files
authored
rename top-level install_requirements.{py,sh,bat} to install_executorch (#7708)
Call it what it does. I did an automated find/replace for the easy cases in scripts/docs and then manually checked the rest.
1 parent 8494b90 commit ee1d7c3

36 files changed

+62
-62
lines changed

.ci/scripts/utils.sh

+3-3
Original file line numberDiff line numberDiff line change
@@ -17,17 +17,17 @@ retry () {
1717
}
1818

1919
clean_executorch_install_folders() {
20-
./install_requirements.sh --clean
20+
./install_executorch.sh --clean
2121
}
2222

2323
install_executorch() {
2424
which pip
2525
# Install executorch, this assumes that Executorch is checked out in the
2626
# current directory.
2727
if [[ "${1:-}" == "use-pt-pinned-commit" ]]; then
28-
./install_requirements.sh --pybind xnnpack --use-pt-pinned-commit
28+
./install_executorch.sh --pybind xnnpack --use-pt-pinned-commit
2929
else
30-
./install_requirements.sh --pybind xnnpack
30+
./install_executorch.sh --pybind xnnpack
3131
fi
3232
# Just print out the list of packages for debugging
3333
pip list

.github/workflows/apple.yml

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ on:
99
paths:
1010
- .ci/scripts/setup-ios.sh
1111
- .github/workflows/apple.yml
12-
- install_requirements.sh
12+
- install_executorch.sh
1313
- backends/apple/**
1414
- build/build_apple_frameworks.sh
1515
- build/build_apple_llm_demo.sh

.github/workflows/pull.yml

+5-5
Original file line numberDiff line numberDiff line change
@@ -200,7 +200,7 @@ jobs:
200200
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
201201
202202
# install pybind
203-
bash install_requirements.sh --pybind xnnpack
203+
bash install_executorch.sh --pybind xnnpack
204204
205205
# install Llava requirements
206206
bash examples/models/llama/install_requirements.sh
@@ -436,7 +436,7 @@ jobs:
436436
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
437437
438438
# install pybind
439-
bash install_requirements.sh --pybind xnnpack
439+
bash install_executorch.sh --pybind xnnpack
440440
441441
# install phi-3-mini requirements
442442
bash examples/models/phi-3-mini/install_requirements.sh
@@ -463,7 +463,7 @@ jobs:
463463
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
464464
465465
# install pybind
466-
bash install_requirements.sh --pybind xnnpack
466+
bash install_executorch.sh --pybind xnnpack
467467
468468
# install llama requirements
469469
bash examples/models/llama/install_requirements.sh
@@ -490,7 +490,7 @@ jobs:
490490
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
491491
492492
# install pybind
493-
bash install_requirements.sh --pybind xnnpack
493+
bash install_executorch.sh --pybind xnnpack
494494
495495
# install llama requirements
496496
bash examples/models/llama/install_requirements.sh
@@ -517,7 +517,7 @@ jobs:
517517
PYTHON_EXECUTABLE=python bash .ci/scripts/setup-linux.sh "cmake"
518518
519519
# install pybind
520-
bash install_requirements.sh --pybind xnnpack
520+
bash install_executorch.sh --pybind xnnpack
521521
522522
# install llama requirements
523523
bash examples/models/llama/install_requirements.sh

backends/apple/mps/setup.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -97,7 +97,7 @@ I 00:00:00.122615 executorch:mps_executor_runner.mm:501] Model verified successf
9797
### [Optional] Run the generated model directly using pybind
9898
1. Make sure `pybind` MPS support was installed:
9999
```bash
100-
./install_requirements.sh --pybind mps
100+
./install_executorch.sh --pybind mps
101101
```
102102
2. Run the `mps_example` script to trace the model and run it directly from python:
103103
```bash

backends/cadence/build_cadence_fusionG3.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ unset XTENSA_CORE
1212
export XTENSA_CORE=FCV_FG3GP
1313
git submodule sync
1414
git submodule update --init
15-
./install_requirements.sh
15+
./install_executorch.sh
1616

1717
rm -rf cmake-out
1818

backends/cadence/build_cadence_hifi4.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ unset XTENSA_CORE
1212
export XTENSA_CORE=nxp_rt600_RI23_11_newlib
1313
git submodule sync
1414
git submodule update --init
15-
./install_requirements.sh
15+
./install_executorch.sh
1616

1717
rm -rf cmake-out
1818

backends/vulkan/docs/android_demo.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ First, build and install ExecuTorch libraries, then build the LLaMA runner
8181
binary using the Android NDK toolchain.
8282

8383
```shell
84-
./install_requirements.sh --clean
84+
./install_executorch.sh --clean
8585
(mkdir cmake-android-out && \
8686
cmake . -DCMAKE_INSTALL_PREFIX=cmake-android-out \
8787
-DCMAKE_TOOLCHAIN_FILE=$ANDROID_NDK/build/cmake/android.toolchain.cmake \

backends/xnnpack/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ After exporting the XNNPACK Delegated model, we can now try running it with exam
9898
cd executorch
9999

100100
# Get a clean cmake-out directory
101-
./install_requirements.sh --clean
101+
./install_executorch.sh --clean
102102
mkdir cmake-out
103103

104104
# Configure cmake

build/test_ios.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ say "Installing Requirements"
6363

6464
pip install --upgrade cmake pip setuptools wheel zstd
6565

66-
./install_requirements.sh --pybind coreml mps xnnpack
66+
./install_executorch.sh --pybind coreml mps xnnpack
6767
export PATH="$(realpath third-party/flatbuffers/cmake-out):$PATH"
6868
./build/install_flatc.sh
6969

docs/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ To build the documentation locally:
6565
1. Run:
6666

6767
```bash
68-
bash install_requirements.sh
68+
bash install_executorch.sh
6969
```
7070

7171
1. Go to the `docs/` directory.

docs/source/apple-runtime.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ python3 -m venv .venv && source .venv/bin/activate && pip install --upgrade pip
109109
4. Install the required dependencies, including those needed for the backends like [Core ML](build-run-coreml.md) or [MPS](build-run-mps.md), if you plan to build them as well:
110110

111111
```bash
112-
./install_requirements.sh --pybind coreml mps xnnpack
112+
./install_executorch.sh --pybind coreml mps xnnpack
113113

114114
# Optional dependencies for Core ML backend.
115115
./backends/apple/coreml/scripts/install_requirements.sh

docs/source/build-run-xtensa.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -162,7 +162,7 @@ In order to run the CMake build, you need the path to the following:
162162

163163
```bash
164164
cd executorch
165-
./install_requirements.sh --clean
165+
./install_executorch.sh --clean
166166
mkdir cmake-out
167167
# prebuild and install executorch library
168168
cmake -DCMAKE_TOOLCHAIN_FILE=<path_to_executorch>/backends/cadence/cadence.cmake \

docs/source/getting-started-setup.md

+10-10
Original file line numberDiff line numberDiff line change
@@ -92,23 +92,23 @@ Alternatively, if you would like to experiment with ExecuTorch quickly and easil
9292
# Install ExecuTorch pip package and its dependencies, as well as
9393
# development tools like CMake.
9494
# If developing on a Mac, make sure to install the Xcode Command Line Tools first.
95-
./install_requirements.sh
95+
./install_executorch.sh
9696
```
9797

98-
Use the [`--pybind` flag](https://github.com/pytorch/executorch/blob/main/install_requirements.sh#L26-L29) to install with pybindings and dependencies for other backends.
98+
Use the [`--pybind` flag](https://github.com/pytorch/executorch/blob/main/install_executorch.sh#L26-L29) to install with pybindings and dependencies for other backends.
9999
```bash
100-
./install_requirements.sh --pybind <coreml | mps | xnnpack>
100+
./install_executorch.sh --pybind <coreml | mps | xnnpack>
101101

102102
# Example: pybindings with CoreML *only*
103-
./install_requirements.sh --pybind coreml
103+
./install_executorch.sh --pybind coreml
104104

105105
# Example: pybinds with CoreML *and* XNNPACK
106-
./install_requirements.sh --pybind coreml xnnpack
106+
./install_executorch.sh --pybind coreml xnnpack
107107
```
108108

109-
By default, `./install_requirements.sh` command installs pybindings for XNNPACK. To disable any pybindings altogether:
109+
By default, `./install_executorch.sh` command installs pybindings for XNNPACK. To disable any pybindings altogether:
110110
```bash
111-
./install_requirements.sh --pybind off
111+
./install_executorch.sh --pybind off
112112
```
113113

114114
After setting up your environment, you are ready to convert your PyTorch programs
@@ -125,7 +125,7 @@ to ExecuTorch.
125125
>
126126
> ```bash
127127
> # From the root of the executorch repo:
128-
> ./install_requirements.sh --clean
128+
> ./install_executorch.sh --clean
129129
> git submodule sync
130130
> git submodule update --init
131131
> ```
@@ -208,7 +208,7 @@ The ExecuTorch repo uses CMake to build its C++ code. Here, we'll configure it t
208208
```bash
209209
# Clean and configure the CMake build system. Compiled programs will
210210
# appear in the executorch/cmake-out directory we create here.
211-
./install_requirements.sh --clean
211+
./install_executorch.sh --clean
212212
(mkdir cmake-out && cd cmake-out && cmake ..)
213213

214214
# Build the executor_runner target
@@ -226,7 +226,7 @@ The ExecuTorch repo uses CMake to build its C++ code. Here, we'll configure it t
226226
>
227227
> ```bash
228228
> # From the root of the executorch repo:
229-
> ./install_requirements.sh --clean
229+
> ./install_executorch.sh --clean
230230
> git submodule sync
231231
> git submodule update --init
232232
> ```

docs/source/llm/getting-started.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ git submodule update --init
5252
# Create a conda environment and install requirements.
5353
conda create -yn executorch python=3.10.0
5454
conda activate executorch
55-
./install_requirements.sh
55+
./install_executorch.sh
5656
5757
cd ../..
5858
```
@@ -83,7 +83,7 @@ cd third-party/executorch
8383
git submodule update --init
8484
8585
# Install requirements.
86-
PYTHON_EXECUTABLE=python ./install_requirements.sh
86+
PYTHON_EXECUTABLE=python ./install_executorch.sh
8787
8888
cd ../..
8989
```
@@ -396,7 +396,7 @@ At this point, the working directory should contain the following files:
396396

397397
If all of these are present, you can now build and run:
398398
```bash
399-
./install_requirements.sh --clean
399+
./install_executorch.sh --clean
400400
(mkdir cmake-out && cd cmake-out && cmake ..)
401401
cmake --build cmake-out -j10
402402
./cmake-out/nanogpt_runner

docs/source/runtime-build-and-cross-compilation.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ cd executorch
4545

4646
# Clean and configure the CMake build system. It's good practice to do this
4747
# whenever cloning or pulling the upstream repo.
48-
./install_requirements.sh --clean
48+
./install_executorch.sh --clean
4949
(mkdir cmake-out && cd cmake-out && cmake ..)
5050
```
5151

@@ -122,7 +122,7 @@ Following are instruction on how to perform cross compilation for Android and iO
122122
Assuming Android NDK is available, run:
123123
```bash
124124
# Run the following lines from the `executorch/` folder
125-
./install_requirements.sh --clean
125+
./install_executorch.sh --clean
126126
mkdir cmake-android-out && cd cmake-android-out
127127

128128
# point -DCMAKE_TOOLCHAIN_FILE to the location where ndk is installed

docs/source/tutorial-xnnpack-delegate-lowering.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -147,7 +147,7 @@ After exporting the XNNPACK Delegated model, we can now try running it with exam
147147
cd executorch
148148

149149
# Get a clean cmake-out directory
150-
./install_requirements.sh --clean
150+
./install_executorch.sh --clean
151151
mkdir cmake-out
152152

153153
# Configure cmake

examples/demo-apps/android/ExecuTorchDemo/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ export ANDROID_NDK=<path-to-android-ndk>
7070
export ANDROID_ABI=arm64-v8a
7171

7272
# Run the following lines from the `executorch/` folder
73-
./install_requirements.sh --clean
73+
./install_executorch.sh --clean
7474
mkdir cmake-android-out
7575

7676
# Build the core executorch library
@@ -114,7 +114,7 @@ export ANDROID_NDK=<path-to-android-ndk>
114114
export ANDROID_ABI=arm64-v8a
115115
export QNN_SDK_ROOT=<path-to-qnn-sdk>
116116

117-
./install_requirements.sh --clean
117+
./install_executorch.sh --clean
118118
mkdir cmake-android-out
119119
cmake . -DCMAKE_INSTALL_PREFIX=cmake-android-out \
120120
-DCMAKE_TOOLCHAIN_FILE="${ANDROID_NDK}/build/cmake/android.toolchain.cmake" \

examples/demo-apps/android/LlamaDemo/docs/delegates/mediatek_README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ git submodule update --init
3636
```
3737
Install dependencies
3838
```
39-
./install_requirements.sh
39+
./install_executorch.sh
4040
```
4141
## Setup Environment Variables
4242
### Download Buck2 and make executable

examples/demo-apps/android/LlamaDemo/docs/delegates/qualcomm_README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ git submodule update --init
3434
```
3535
Install dependencies
3636
```
37-
./install_requirements.sh
37+
./install_executorch.sh
3838
```
3939

4040
## Setup QNN

examples/demo-apps/android/LlamaDemo/docs/delegates/xnnpack_README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -36,12 +36,12 @@ git submodule update --init
3636
```
3737
Install dependencies
3838
```
39-
./install_requirements.sh
39+
./install_executorch.sh
4040
```
4141

4242
Optional: Use the --pybind flag to install with pybindings.
4343
```
44-
./install_requirements.sh --pybind xnnpack
44+
./install_executorch.sh --pybind xnnpack
4545
```
4646

4747

examples/demo-apps/apple_ios/ExecuTorchDemo/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ python3 -m venv .venv && source .venv/bin/activate
5151

5252
pip install --upgrade cmake pip setuptools wheel
5353

54-
./install_requirements.sh --pybind coreml mps xnnpack
54+
./install_executorch.sh --pybind coreml mps xnnpack
5555
```
5656

5757
### 4. Backend Dependencies

examples/demo-apps/apple_ios/LLaMA/docs/delegates/mps_README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ git submodule update --init
3333
Install dependencies
3434

3535
```
36-
./install_requirements.sh
36+
./install_executorch.sh
3737
```
3838

3939
## Prepare Models

examples/demo-apps/apple_ios/LLaMA/docs/delegates/xnnpack_README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -32,11 +32,11 @@ git submodule update --init
3232
Install dependencies
3333

3434
```
35-
./install_requirements.sh
35+
./install_executorch.sh
3636
```
3737
Optional: Use the --pybind flag to install with pybindings.
3838
```
39-
./install_requirements.sh --pybind xnnpack
39+
./install_executorch.sh --pybind xnnpack
4040
```
4141
## Prepare Models
4242
In this demo app, we support text-only inference with up-to-date Llama models and image reasoning inference with LLaVA 1.5.

examples/demo-apps/react-native/rnllama/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ A React Native mobile application for running LLaMA language models using ExecuT
2626

2727
3. Pull submodules: `git submodule sync && git submodule update --init`
2828

29-
4. Install dependencies: `./install_requirements.sh --pybind xnnpack && ./examples/models/llama/install_requirements.sh`
29+
4. Install dependencies: `./install_executorch.sh --pybind xnnpack && ./examples/models/llama/install_requirements.sh`
3030

3131
5. Follow the instructions in the [README](https://github.com/pytorch/executorch/blob/main/examples/models/llama/README.md#option-a-download-and-export-llama32-1b3b-model) to export a model as `.pte`
3232

@@ -40,4 +40,4 @@ A React Native mobile application for running LLaMA language models using ExecuT
4040

4141
10. Select the model and tokenizer in the app to start chatting:
4242

43-
[![rnllama]](https://github.com/user-attachments/assets/b339f1ec-8b80-41f0-b3f6-ded6698ac926)
43+
[![rnllama]](https://github.com/user-attachments/assets/b339f1ec-8b80-41f0-b3f6-ded6698ac926)

examples/devtools/build_example_runner.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ done
3737
main() {
3838
cd "${EXECUTORCH_ROOT}"
3939

40-
./install_requirements.sh --clean
40+
./install_executorch.sh --clean
4141

4242
if [[ "${BUILD_COREML}" == "ON" ]]; then
4343
cmake -DCMAKE_INSTALL_PREFIX=cmake-out \

examples/models/llama/README.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -148,7 +148,7 @@ Llama 3 8B performance was measured on the Samsung Galaxy S22, S24, and OnePlus
148148
## Step 1: Setup
149149
> :warning: **double check your python environment**: make sure `conda activate <VENV>` is run before all the bash and python scripts.
150150
151-
1. Follow the [tutorial](https://pytorch.org/executorch/main/getting-started-setup) to set up ExecuTorch. For installation run `./install_requirements.sh --pybind xnnpack`
151+
1. Follow the [tutorial](https://pytorch.org/executorch/main/getting-started-setup) to set up ExecuTorch. For installation run `./install_executorch.sh --pybind xnnpack`
152152
2. Run `examples/models/llama/install_requirements.sh` to install a few dependencies.
153153

154154

@@ -440,8 +440,8 @@ This example tries to reuse the Python code, with minimal modifications to make
440440
```
441441
git clean -xfd
442442
pip uninstall executorch
443-
./install_requirements.sh --clean
444-
./install_requirements.sh --pybind xnnpack
443+
./install_executorch.sh --clean
444+
./install_executorch.sh --pybind xnnpack
445445
```
446446
- If you encounter `pthread` related issues during link time, add `pthread` in `target_link_libraries` in `CMakeLists.txt`
447447
- On Mac, if there is linking error in Step 4 with error message like

0 commit comments

Comments
 (0)