You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> **Note:** More details about supported models for vLLM-IPEX-XPU can be found at [supported-models](https://github.com/intel/llm-scaler/tree/main/vllm#3-supported-models).
38
41
39
42
### System Requirements for LLM Models
40
43
@@ -70,6 +73,7 @@ In this microservices, we have supported following backend LLM service as integr
70
73
-[Bedrock](./README_bedrock.md)
71
74
-[Native](./README_native.md), based on optimum habana
72
75
-[Predictionguard](./README_predictionguard.md)
76
+
-[VLLM-IPEX](./README_vllm_ipex.md), based on B60 Graphics
2. Navigate to the Docker Compose directory and start the services:
58
+
```bash
59
+
cd comps/llms/deployment/docker_compose/
60
+
docker compose -f compose_text-generation.yaml up textgen-vllm-ipex-service -d
61
+
```
62
+
63
+
>**Note:** More details about supported models can be found at [supported-models](https://github.com/intel/llm-scaler/tree/main/vllm#3-supported-models).
64
+
65
+
---
66
+
67
+
## Consume LLM Service
68
+
69
+
Once the service is running, you can send requests to the API.
Copy file name to clipboardExpand all lines: comps/lvms/README.md
+9-8Lines changed: 9 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -38,11 +38,12 @@ Users can configure and deploy LVM-related services based on their specific requ
38
38
39
39
The LVM Microservice supports multiple implementation options. Select the one that best fits your use case and follow the linked documentation for detailed setup instructions.
2. Navigate to the Docker Compose directory and start the services:
60
+
```bash
61
+
cd comps/lvms/deployment/docker_compose/
62
+
docker compose up lvm-vllm-ipex-service -d
63
+
```
64
+
65
+
>**Note:** More details about supported models can be found at [supported-models](https://github.com/intel/llm-scaler/tree/main/vllm#3-supported-models).
66
+
67
+
---
68
+
69
+
## Consume LVM Service
70
+
71
+
Once the service is running, you can send requests to the API.
72
+
73
+
### Use the LVM Service API
74
+
75
+
Send a POST request with an image url and a prompt.
0 commit comments