Commit 48e11be
committed
server: OpenAI API and llama.cpp Web UI compatibility
- OpenAI-style /v1/models, /v1/chat/completions (SSE), error bodies, CORS
- GET /props and enriched model rows for ggml Web UI
- Session file persistence and optional TILES_BOOTSTRAP_* startup load
- POST /models/load when TILES_MODEL_CACHE_PATH is set (Web UI load)
- Chat SSE: timings (Web UI) and metrics (Tiles CLI memory mode)
- Pytest coverage; TILES_SKIP_SESSION_PERSIST in tests
- just webui-llamacpp via scripts/phase2_llamacpp_webui.sh
Made-with: Cursor1 parent 1b8be89 commit 48e11be
File tree
12 files changed
+1119
-158
lines changed- scripts
- server
- backend
- tests
12 files changed
+1119
-158
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
15 | 15 | | |
16 | 16 | | |
17 | 17 | | |
| 18 | + | |
| 19 | + | |
| 20 | + | |
| 21 | + | |
| 22 | + | |
| 23 | + | |
| 24 | + | |
| 25 | + | |
18 | 26 | | |
19 | 27 | | |
20 | 28 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
| 1 | + | |
| 2 | + | |
| 3 | + | |
| 4 | + | |
| 5 | + | |
| 6 | + | |
| 7 | + | |
| 8 | + | |
| 9 | + | |
| 10 | + | |
| 11 | + | |
| 12 | + | |
| 13 | + | |
| 14 | + | |
| 15 | + | |
| 16 | + | |
| 17 | + | |
| 18 | + | |
| 19 | + | |
| 20 | + | |
| 21 | + | |
| 22 | + | |
| 23 | + | |
| 24 | + | |
| 25 | + | |
| 26 | + | |
| 27 | + | |
| 28 | + | |
| 29 | + | |
| 30 | + | |
| 31 | + | |
| 32 | + | |
| 33 | + | |
| 34 | + | |
| 35 | + | |
| 36 | + | |
| 37 | + | |
| 38 | + | |
| 39 | + | |
| 40 | + | |
| 41 | + | |
| 42 | + | |
| 43 | + | |
0 commit comments