Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
17 changes: 17 additions & 0 deletions services/ollama/.env
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
#version=1.1
#URL=https://github.com/tailscale-dev/ScaleTail
#COMPOSE_PROJECT_NAME= # Optional: only use when running multiple deployments on the same infrastructure.

# Service Configuration
SERVICE=ollama
IMAGE_URL=ollama/ollama:latest

# Network Configuration
SERVICEPORT=11434 # Ollama's default API port. Uncomment the "ports:" section in compose.yaml to expose to LAN.
DNS_SERVER=9.9.9.9 # Preferred DNS server for Tailscale. Uncomment the "dns:" section in compose.yaml to enable.

# Tailscale Configuration
TS_AUTHKEY= # Auth key from https://tailscale.com/admin/authkeys. See: https://tailscale.com/kb/1085/auth-keys#generate-an-auth-key for instructions.

# Ollama-specific variables
OLLAMA_API_KEY= # Optional: set a secret key to restrict API access (leave blank to disable auth)
101 changes: 101 additions & 0 deletions services/ollama/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,101 @@
# Ollama with Tailscale Sidecar Configuration

This Docker Compose configuration sets up [Ollama](https://ollama.com) with Tailscale as a sidecar container to keep the API reachable securely over your Tailnet.

## Ollama

[Ollama](https://ollama.com) lets you run large language models (LLMs) locally — such as Llama 3, Mistral, and Gemma — with a simple API compatible with the OpenAI client format. Pairing it with Tailscale means you can access your local models from any device on your Tailnet (phone, laptop, remote machine) without exposing the API to the public internet.

## Configuration Overview

In this setup, the `tailscale-ollama` service runs Tailscale, which manages secure networking for Ollama. The `app-ollama` service uses Docker's `network_mode: service:tailscale` so all traffic is routed through the Tailscale network stack. The Ollama API remains Tailnet-only by default unless you explicitly expose the port to your LAN.

An optional `yourNetwork` external Docker network is attached to the `tailscale` container. This allows other containers on the same host (such as Open WebUI or other LLM frontends) to reach Ollama via its Tailscale IP, keeping inter-container communication on the same overlay network.

## Prerequisites

- The host user must be in the `docker` group.
- The `/dev/net/tun` device must be available on the host (standard on most Linux systems).
- Pre-create the bind-mount directories before starting the stack to avoid Docker creating root-owned folders:

```bash
mkdir -p config ts/state ollama-data
```

- If you use the optional `yourNetwork` network, create it first if it does not already exist:

```bash
docker network create yourNetwork
```

If you don't use a shared proxy network, remove the `networks:` sections from `compose.yaml`.

## Volumes

| Path | Purpose |
|------|---------|
| `./config` | Tailscale serve config (`serve.json`) |
| `./ts/state` | Tailscale persistent state |
| `./ollama-data` | Downloaded Ollama models (can be large — ensure enough disk space) |

## MagicDNS and HTTPS

Tailscale Serve is pre-configured to proxy HTTPS on port 443 to Ollama's internal port 11434. To enable it:

1. Uncomment `TS_ACCEPT_DNS=true` in the `tailscale` service environment.
2. Ensure your Tailnet has MagicDNS and HTTPS certificates enabled in the [Tailscale admin console](https://login.tailscale.com/admin/dns).
3. The `serve.json` config in `compose.yaml` uses `$TS_CERT_DOMAIN` automatically — no manual editing needed.

You can then reach Ollama at `https://ollama.<your-tailnet-name>.ts.net`.

## Port Exposure (LAN access)

By default, the `ports:` section is commented out — Ollama is only accessible over your Tailnet. If you also want LAN access (e.g. from devices not on Tailscale), uncomment it in `compose.yaml`:

```yaml
ports:
- 0.0.0.0:11434:11434
```

This is optional and not required for Tailnet-only usage.

## API Key (Optional)

Ollama supports a simple bearer token for API access. Set `OLLAMA_API_KEY` in your `.env` file to enable it. Leave it blank to allow unauthenticated access (safe when Tailnet-only).

## First-time Setup

After starting the stack, pull a model to get started:

```bash
docker exec app-ollama ollama pull llama3
```

You can then send requests to the API:

```bash
curl http://<tailscale-ip>:11434/api/generate \
-d '{"model": "llama3", "prompt": "Hello!"}'
```

Or if using HTTPS via Tailscale Serve:

```bash
curl https://ollama.<your-tailnet-name>.ts.net/api/generate \
-d '{"model": "llama3", "prompt": "Hello!"}'
```

## Files to check

Please check the following contents for validity as some variables need to be defined upfront.

- `.env` — Set `TS_AUTHKEY` (required). Optionally set `OLLAMA_API_KEY`.

## Useful Links

- [Ollama official site](https://ollama.com)
- [Ollama model library](https://ollama.com/library)
- [Ollama GitHub](https://github.com/ollama/ollama)
- [Tailscale auth keys](https://tailscale.com/kb/1085/auth-keys)
- [Tailscale Serve docs](https://tailscale.com/kb/1312/serve)
- [Open WebUI](https://github.com/open-webui/open-webui) — a popular browser-based UI for Ollama
78 changes: 78 additions & 0 deletions services/ollama/compose.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,78 @@
configs:
ts-serve:
content: |
{"TCP":{"443":{"HTTPS":true}},
"Web":{"$${TS_CERT_DOMAIN}:443":
{"Handlers":{"/":
{"Proxy":"http://127.0.0.1:11434"}}}},
"AllowFunnel":{"$${TS_CERT_DOMAIN}:443":false}}

services:
# Make sure you have updated/checked the .env file with the correct variables.
# All the ${ xx } need to be defined there.

# Tailscale Sidecar Configuration
tailscale:
image: tailscale/tailscale:latest # Image to be used
container_name: tailscale-${SERVICE} # Name for local container management
hostname: ${SERVICE} # Name used within your Tailscale environment
environment:
- TS_AUTHKEY=${TS_AUTHKEY}
- TS_STATE_DIR=/var/lib/tailscale
- TS_SERVE_CONFIG=/config/serve.json # Tailscale Serve configuration to expose the web interface on your local Tailnet - remove this line if not required
- TS_USERSPACE=false
- TS_ENABLE_HEALTH_CHECK=true # Enable healthcheck endpoint: "/healthz"
- TS_LOCAL_ADDR_PORT=127.0.0.1:41234 # The <addr>:<port> for the healthz endpoint
#- TS_ACCEPT_DNS=true # Uncomment when using MagicDNS
- TS_AUTH_ONCE=true
configs:
- source: ts-serve
target: /config/serve.json
volumes:
- ./config:/config # Config folder used to store Tailscale files
- ./ts/state:/var/lib/tailscale # Tailscale requirement
devices:
- /dev/net/tun:/dev/net/tun # Network configuration for Tailscale to work
cap_add:
- net_admin # Tailscale requirement
- sys_module # Required to load kernel modules for Tailscale
#ports:
# - 0.0.0.0:${SERVICEPORT}:${SERVICEPORT} # Binding port ${SERVICE}PORT to the local network - may be removed if only exposure to your Tailnet is required
# If any DNS issues arise, use your preferred DNS provider by uncommenting the config below
#dns:
# - ${DNS_SERVER}
networks:
- yourNetwork # Optional: connect to an existing proxy network so other containers can reach Ollama via its Tailscale IP
healthcheck:
test: ["CMD", "wget", "--spider", "-q", "http://127.0.0.1:41234/healthz"] # Check Tailscale has a Tailnet IP and is operational
interval: 1m # How often to perform the check
timeout: 10s # Time to wait for the check to succeed
retries: 3 # Number of retries before marking as unhealthy
start_period: 10s # Time to wait before starting health checks
restart: always

# Ollama
application:
image: ${IMAGE_URL} # Image to be used
network_mode: service:tailscale # Sidecar configuration to route Ollama through Tailscale
container_name: app-${SERVICE} # Name for local container management
environment:
- OLLAMA_HOST=0.0.0.0:11434
- OLLAMA_API_KEY=${OLLAMA_API_KEY} # Optional: set an API key to restrict access
- OLLAMA_KEEP_ALIVE=24h # Optional: keeps models loaded in memory (default is 5 min)
volumes:
- ./${SERVICE}-data:/root/.ollama # Stores downloaded models
depends_on:
tailscale:
condition: service_healthy
healthcheck:
test: ["CMD", "pgrep", "-f", "${SERVICE}"] # Check if Ollama process is running
interval: 1m # How often to perform the check
timeout: 10s # Time to wait for the check to succeed
retries: 3 # Number of retries before marking as unhealthy
start_period: 30s # Time to wait before starting health checks
restart: always

networks:
yourNetwork:
external: true # Assumes an existing external Docker network named "yourNetwork"