Skip to content

Commit 6bdb1ba

Browse files
authored
Rename vllm dockerfile (#3330)
1 parent c161926 commit 6bdb1ba

File tree

6 files changed

+7
-7
lines changed

6 files changed

+7
-7
lines changed

README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -85,9 +85,9 @@ curl -X POST -d '{"prompt":"count from 1 to 9 in french ", "max_tokens": 100}' -
8585

8686
```bash
8787
#export token=<HUGGINGFACE_HUB_TOKEN>
88-
docker build --pull . -f docker/Dockerfile.llm -t ts/llm
88+
docker build --pull . -f docker/Dockerfile.vllm -t ts/vllm
8989

90-
docker run --rm -ti --shm-size 10g --gpus all -e HUGGING_FACE_HUB_TOKEN=$token -p 8080:8080 -v data:/data ts/llm --model_id meta-llama/Meta-Llama-3-8B-Instruct --disable_token_auth
90+
docker run --rm -ti --shm-size 10g --gpus all -e HUGGING_FACE_HUB_TOKEN=$token -p 8080:8080 -v data:/data ts/vllm --model_id meta-llama/Meta-Llama-3-8B-Instruct --disable_token_auth
9191

9292
# Try it out
9393
curl -X POST -d '{"model":"meta-llama/Meta-Llama-3-8B-Instruct", "prompt":"Hello, my name is", "max_tokens": 200}' --header "Content-Type: application/json" "http://localhost:8080/predictions/model/1.0/v1/completions"
File renamed without changes.

docs/llm_deployment.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ The launcher can either be used standalone or in combination with our provided T
1111

1212
To launch the docker we first need to build it:
1313
```bash
14-
docker build . -f docker/Dockerfile.llm -t ts/llm
14+
docker build . -f docker/Dockerfile.vllm -t ts/vllm
1515
```
1616

1717
Models are usually loaded from the HuggingFace hub and are cached in a [docker volume](https://docs.docker.com/storage/volumes/) for faster reload.
@@ -22,7 +22,7 @@ export token=<HUGGINGFACE_HUB_TOKEN>
2222

2323
You can then go ahead and launch a TorchServe instance serving your selected model:
2424
```bash
25-
docker run --rm -ti --shm-size 1g --gpus all -e HUGGING_FACE_HUB_TOKEN=$token -p 8080:8080 -v data:/data ts/llm --model_id meta-llama/Meta-Llama-3-8B-Instruct --disable_token_auth
25+
docker run --rm -ti --shm-size 1g --gpus all -e HUGGING_FACE_HUB_TOKEN=$token -p 8080:8080 -v data:/data ts/vllm --model_id meta-llama/Meta-Llama-3-8B-Instruct --disable_token_auth
2626
```
2727

2828
To change the model you just need to exchange the identifier given to the `--model_id` parameter.

examples/large_models/vllm/llama3/Readme.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ To leverage the power of vLLM we fist need to install it using pip in out develo
99
```bash
1010
python -m pip install -r ../requirements.txt
1111
```
12-
For later deployments we can make vLLM part of the deployment environment by adding the requirements.txt while building the model archive in step 2 (see [here](../../../../model-archiver/README.md#model-specific-custom-python-requirements) for details) or we can make it part of a docker image like [here](../../../../docker/Dockerfile.llm).
12+
For later deployments we can make vLLM part of the deployment environment by adding the requirements.txt while building the model archive in step 2 (see [here](../../../../model-archiver/README.md#model-specific-custom-python-requirements) for details) or we can make it part of a docker image like [here](../../../../docker/Dockerfile.vllm).
1313

1414
### Step 1: Download Model from HuggingFace
1515

examples/large_models/vllm/lora/Readme.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ To leverage the power of vLLM we fist need to install it using pip in out develo
99
```bash
1010
python -m pip install -r ../requirements.txt
1111
```
12-
For later deployments we can make vLLM part of the deployment environment by adding the requirements.txt while building the model archive in step 2 (see [here](../../../../model-archiver/README.md#model-specific-custom-python-requirements) for details) or we can make it part of a docker image like [here](../../../../docker/Dockerfile.llm).
12+
For later deployments we can make vLLM part of the deployment environment by adding the requirements.txt while building the model archive in step 2 (see [here](../../../../model-archiver/README.md#model-specific-custom-python-requirements) for details) or we can make it part of a docker image like [here](../../../../docker/Dockerfile.vllm).
1313

1414
### Step 1: Download Model from HuggingFace
1515

examples/large_models/vllm/mistral/Readme.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ To leverage the power of vLLM we fist need to install it using pip in out develo
99
```bash
1010
python -m pip install -r ../requirements.txt
1111
```
12-
For later deployments we can make vLLM part of the deployment environment by adding the requirements.txt while building the model archive in step 2 (see [here](../../../../model-archiver/README.md#model-specific-custom-python-requirements) for details) or we can make it part of a docker image like [here](../../../../docker/Dockerfile.llm).
12+
For later deployments we can make vLLM part of the deployment environment by adding the requirements.txt while building the model archive in step 2 (see [here](../../../../model-archiver/README.md#model-specific-custom-python-requirements) for details) or we can make it part of a docker image like [here](../../../../docker/Dockerfile.vllm).
1313

1414
### Step 1: Download Model from HuggingFace
1515

0 commit comments

Comments
 (0)