-
Notifications
You must be signed in to change notification settings - Fork 162
Expand file tree
/
Copy pathdemo.py
More file actions
51 lines (45 loc) · 1.62 KB
/
demo.py
File metadata and controls
51 lines (45 loc) · 1.62 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
# ---------------------------------------------------------------------
# Copyright (c) 2025 Qualcomm Technologies, Inc. and/or its subsidiaries.
# SPDX-License-Identifier: BSD-3-Clause
# ---------------------------------------------------------------------
from __future__ import annotations
from qai_hub_models.models._shared.llama3.model import END_TOKENS
from qai_hub_models.models._shared.llm.demo import llm_chat_demo
from qai_hub_models.models._shared.llm.model import LLM_QNN, LLM_AIMETOnnx, LLMBase
from qai_hub_models.models.llama_v3_8b_instruct import (
MODEL_ID,
FP_Model,
Model,
QNN_Model,
)
from qai_hub_models.models.llama_v3_8b_instruct.model import (
HF_REPO_NAME,
HF_REPO_URL,
SUPPORTED_PRECISIONS,
)
from qai_hub_models.utils.checkpoint import CheckpointSpec
def llama_3_chat_demo(
model_cls: type[LLM_AIMETOnnx] = Model,
fp_model_cls: type[LLMBase] = FP_Model,
qnn_model_cls: type[LLM_QNN] = QNN_Model,
model_id: str = MODEL_ID,
end_tokens: set = END_TOKENS,
hf_repo_name: str = HF_REPO_NAME,
hf_repo_url: str = HF_REPO_URL,
default_prompt: str | None = None,
test_checkpoint: CheckpointSpec | None = None,
) -> None:
llm_chat_demo(
model_cls=model_cls,
fp_model_cls=fp_model_cls,
qnn_model_cls=qnn_model_cls,
model_id=model_id,
end_tokens=end_tokens,
hf_repo_name=hf_repo_name,
hf_repo_url=hf_repo_url,
supported_precisions=SUPPORTED_PRECISIONS,
default_prompt=default_prompt,
test_checkpoint=test_checkpoint,
)
if __name__ == "__main__":
llama_3_chat_demo(model_cls=Model)