- Overview
- Features
- Setup Instructions
- Environment Setup
- LLM Suite
- Set up QNN SDK (Qualcomm AI stack)
- Run the Application
This demo application demonstrates how to integrate a Large Language Model (LLM) locally. It enables users to query aspects using natural language and receive contextually enriched responses.
- Task customization enabled through configurable prompt file.
- Support for multiple languages.
- Support for multiple LLM from Hugging face.
- Choose between various models.
Before proceeding further, please ensure that all the setup steps outlined below are completed in the specified order. These instructions are essential for configuring the various tools required to successfully run the application.
Each section provides a reference to internal documentation for detailed guidance. Please follow them carefully to avoid any setup issues later in the process.
Miniconda is required to manage the application's Python environment and dependencies. Please follow the setup instructions carefully to ensure a consistent and reproducible environment.
For detailed steps, refer to the internal documentation: Set up Miniconda
Git is required for version control and collaboration. Proper configuration ensures seamless integration with repositories and development workflows.
For detailed steps, refer to the internal documentation: Setup Git
To set up the Python environment required for running the application, follow the steps below. This ensures all dependencies are installed in an isolated and reproducible environment.
-
Create your working directory :
mkdir my_working_directory cd my_working_directory -
Download Your Application :
git clone -n --depth=1 --filter=tree:0 https://github.com/qualcomm/Startup-Demos.git cd Startup-Demos git sparse-checkout set --no-cone /GenAI/AI_PC/History_LLM git checkout
-
Navigate to Application Directory :
cd ./GenAI/AI_PC/History_LLM -
Create a new Conda environment with Python 3.12:
conda create -n myenv python=3.12
-
Activate the environment:
conda activate myenv
-
Install the required dependencies:
pip install -r requirements.txt
💡 Make sure you have Miniconda or Anaconda installed before running these commands.
You can use openly available Large Language Models (LLMs) like locally in LM Studio.
To retrieve your API key for the Cloud AI Playground, refer to the documentation: Download and Install
To retrieve your API key for the Cloud AI Playground, refer to the documentation: Download_Models
Once you have obtained the keys, make sure to securely store them in your environment configuration or secrets manager as per your project’s security guidelines.
After selecting the provider, choose the specific model (e.g., Llama-3.3-70B, Llama-3.1-8B, etc.) that best suits your use case: Select_Models
To retrieve your API key for the Cloud AI Playground, refer to the documentation:Load_models
Enter the API key associated with the selected LLM instance. This is required to authenticate and access the model's capabilities: Start_Local_server
Once you have obtained the keys, make sure to securely store them in your environment configuration or secrets manager as per your project’s security guidelines.
The Qualcomm AI Engine Direct SDK is a software development kit that enables developers to efficiently deploy and optimize AI models directly on Qualcomm® Snapdragon™ platforms, leveraging on-device acceleration for enhanced performance and power efficiency.
For instructions on setting environment variables on your system, refer to the documentation: QNN_SDK
💡 Once set, check your application can access API server in your functions.
To launch the application, execute the following command from the application directory:
streamlit run .\src\ChatApp.pyThis will start the interactive web interface where you can configure and use the application.
Prepare the context prompt file with keywords that specify all user inputs you want the application to process. This file will be parsed and embedded for querying.
streamlit run .\src\Prompts_Utils\prompt_editor.pyDecide how you want to manage the Query data in the database. The following options are available:
- Use Existing – Query the existing database without modifying it.
- Add development aspects – Append some development acpects to the current database to make it perticular.
- Select the mode/Topic – Select the perticular Topic.
- Select the region – Select the geographical regions.
- Select the language – Select the language.
- Select the respose – Select the response length based on the requirement.
✅ Once all configurations are complete, you can begin interacting with the application through the chat interface.



