This page describes how to run OpenChat Playground (OCP) with Hugging Face models integration.
-
Get the repository root.
# bash/zsh REPOSITORY_ROOT=$(git rev-parse --show-toplevel)
# PowerShell $REPOSITORY_ROOT = git rev-parse --show-toplevel
-
Make sure the Ollama server is up and running.
ollama serve
-
Download the Hugging Face model. The default model OCP uses is
Qwen/Qwen3-0.6B-GGUF.ollama pull hf.co/Qwen/Qwen3-0.6B-GGUF
Alternatively, if you want to run with a different model, say microsoft/phi-4-gguf, other than the default one, download it first by running the following command.
ollama pull hf.co/microsoft/phi-4-gguf
Make sure to follow the exact format like
hf.co/{{org}}/{{model}}and the model MUST includeGGUF. -
Make sure you are at the repository root.
cd $REPOSITORY_ROOT
-
Run the app.
# bash/zsh dotnet run --project $REPOSITORY_ROOT/src/OpenChat.PlaygroundApp -- \ --connector-type HuggingFace
# PowerShell dotnet run --project $REPOSITORY_ROOT\src\OpenChat.PlaygroundApp -- ` --connector-type HuggingFace
Alternatively, if you want to run with a different model, say microsoft/phi-4-gguf, make sure you've already downloaded the model by running the
ollama pull hf.co/microsoft/phi-4-ggufcommand.# bash/zsh dotnet run --project $REPOSITORY_ROOT/src/OpenChat.PlaygroundApp -- \ --connector-type HuggingFace \ --model hf.co/microsoft/phi-4-gguf
# PowerShell dotnet run --project $REPOSITORY_ROOT\src\OpenChat.PlaygroundApp -- ` --connector-type HuggingFace ` --model hf.co/microsoft/phi-4-gguf
-
Open your web browser, navigate to
http://localhost:5280, and enter prompts.
-
Make sure the Ollama server is up and running.
ollama serve
-
Download the Hugging Face model. The default model OCP uses is Qwen/Qwen3-0.6B-GGUF.
ollama pull hf.co/Qwen/Qwen3-0.6B-GGUF
Alternatively, if you want to run with a different model, say microsoft/phi-4-gguf, other than the default one, download it first by running the following command.
ollama pull hf.co/microsoft/phi-4-gguf
Make sure to follow the exact format like
hf.co/{{org}}/{{model}}and the model MUST includeGGUF. -
Make sure you are at the repository root.
cd $REPOSITORY_ROOT
-
Build a container.
docker build -f Dockerfile -t openchat-playground:latest . -
Run the app. The default model OCP uses is Qwen/Qwen3-0.6B-GGUF.
# bash/zsh - from locally built container docker run -i --rm -p 8080:8080 openchat-playground:latest \ --connector-type HuggingFace \ --base-url http://host.docker.internal:11434# PowerShell - from locally built container docker run -i --rm -p 8080:8080 openchat-playground:latest ` --connector-type HuggingFace ` --base-url http://host.docker.internal:11434
# bash/zsh - from GitHub Container Registry docker run -i --rm -p 8080:8080 ghcr.io/aliencube/open-chat-playground/openchat-playground:latest \ --connector-type HuggingFace \ --base-url http://host.docker.internal:11434# PowerShell - from GitHub Container Registry docker run -i --rm -p 8080:8080 ghcr.io/aliencube/open-chat-playground/openchat-playground:latest ` --connector-type HuggingFace ` --base-url http://host.docker.internal:11434
Alternatively, if you want to run with a different model, say microsoft/phi-4-gguf, make sure you've already downloaded the model by running the
ollama pull hf.co/microsoft/phi-4-ggufcommand.# bash/zsh - from locally built container docker run -i --rm -p 8080:8080 openchat-playground:latest \ --connector-type HuggingFace \ --base-url http://host.docker.internal:11434 \ --model hf.co/microsoft/phi-4-gguf# PowerShell - from locally built container docker run -i --rm -p 8080:8080 openchat-playground:latest ` --connector-type HuggingFace ` --base-url http://host.docker.internal:11434 ` --model hf.co/microsoft/phi-4-gguf
# bash/zsh - from GitHub Container Registry docker run -i --rm -p 8080:8080 ghcr.io/aliencube/open-chat-playground/openchat-playground:latest \ --connector-type HuggingFace \ --base-url http://host.docker.internal:11434 \ --model hf.co/microsoft/phi-4-gguf# PowerShell - from GitHub Container Registry docker run -i --rm -p 8080:8080 ghcr.io/aliencube/open-chat-playground/openchat-playground:latest ` --connector-type HuggingFace ` --base-url http://host.docker.internal:11434 ` --model hf.co/microsoft/phi-4-gguf
-
Open your web browser, navigate to
http://localhost:8080, and enter prompts.
-
Make sure you are at the repository root.
cd $REPOSITORY_ROOT
-
Login to Azure.
# Login to Azure Dev CLI azd auth login -
Check login status.
# Azure Dev CLI azd auth login --check-status -
Initialize
azdtemplate.azd init
NOTE: You will be asked to provide environment name for provisioning.
-
Set the connector type to
HuggingFace.azd env set CONNECTOR_TYPE "HuggingFace"
The default model OCP uses is Qwen/Qwen3-0.6B-GGUF. If you want to run with a different model, say microsoft/phi-4-gguf, other than the default one, add it to azd environment variables.
azd env set HUGGING_FACE_MODEL "hf.co/microsoft/phi-4-gguf"
Make sure to follow the exact format like
hf.co/{{org}}/{{model}}and the model MUST includeGGUF. -
As a default, the app uses a Serverless GPU with NVIDIA T4 (
NC8as-T4). If you want to use NVIDIA A100, set the GPU profile.azd env set GPU_PROFILE_NAME "NC24-A100"
If you want to know more about Serverless GPU, visit Using serverless GPUs in Azure Container Apps.
-
Run the following commands in order to provision and deploy the app.
azd up
NOTE: You will be asked to provide Azure subscription and location for deployment. IMPORTANT: Due to the limitation for GPU support, the available regions are limited to
Australia East,Sweden CentralandWest US 3. For more details, visit Using serverless GPUs in Azure Container Apps.Once deployed, you will be able to see the deployed OCP app URL.
-
Open your web browser, navigate to the OCP app URL, and enter prompts.
-
Clean up all the resources.
azd down --force --purge