Skip to content

Commit ffd46e5

Browse files
committed
Remove HFTGI and Flan references
1 parent 1163e61 commit ffd46e5

File tree

6 files changed

+2
-156
lines changed

6 files changed

+2
-156
lines changed

bootstrap/ic-shared-llm/deployment-hftgi.yaml

Lines changed: 0 additions & 109 deletions
This file was deleted.

bootstrap/ic-shared-llm/kustomization.yaml

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,9 +17,6 @@ resources:
1717
# wave 2
1818
- inference-service-granite-modelcar.yaml
1919
- inference-service-qwen-modelcar.yaml
20-
- pvc-hftgi.yaml
21-
- deployment-hftgi.yaml
22-
- service-hftgi.yaml
2320

2421
transformers:
2522
- namespace-transformer.yaml

bootstrap/ic-shared-llm/pvc-hftgi.yaml

Lines changed: 0 additions & 18 deletions
This file was deleted.

bootstrap/ic-shared-llm/service-hftgi.yaml

Lines changed: 0 additions & 24 deletions
This file was deleted.

content/modules/ROOT/pages/02-05-validating-env.adoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ Success: Minio is reachable on minio.ic-shared-minio.svc.cluster.local:9000
3030
Success: Gitea is reachable on gitea.gitea.svc.cluster.local:3000
3131
Success: Postgres Database is reachable on claimdb.ic-shared-db.svc.cluster.local:5432
3232
Success: LLM Service is reachable on llm.ic-shared-llm.svc.cluster.local:8000
33-
Success: LLM Service-FlanT5 is reachable on llm-flant5.ic-shared-llm.svc.cluster.local:3000
33+
Success: LLM Service-Qwen2.5 is reachable on qwen-predictor.ic-shared-llm.svc.cluster.local:8080
3434
Success: ModelMesh is reachable on modelmesh-serving.ic-shared-img-det.svc.cluster.local:8033
3535
Success: Milvus Vector DB is reachable on vectordb-milvus.ic-shared-milvus.svc.cluster.local:19530
3636
----

content/modules/ROOT/pages/03-04-comparing-model-servers.adoc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ include::_attributes.adoc[]
33

44
So far, for this {ic-lab}, we have used the model https://huggingface.co/RedHatAI/granite-3.1-8b-instruct[Granite 3.1 8B Instruct,window=_blank]. Although lighter than other models, it is still quite heavy and we need a large GPU to run it. Would we get as good results with a smaller model running on a CPU only? Let's try!
55
6-
In this exercise, we'll pitch our previous model against a much smaller LLM called https://huggingface.co/google/flan-t5-large[flan-t5-large,window=_blank]. We'll compare the results and see if the smaller model is good enough for our use case.
6+
In this exercise, we'll pitch our previous model against a much smaller LLM called https://huggingface.co/RedHatAI/Qwen2.5-0.5B-quantized.w8a8[Qwen2.5 0.5B Quantized-w8a8,window=_blank]. We'll compare the results and see if the smaller model is good enough for our use case.
77
88
From the `parasol-insurance/lab-materials/03` folder, please open the notebook called `03-04-comparing-model-servers.ipynb` and follow the instructions.
99

0 commit comments

Comments
 (0)