You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
fix: stop sequence truncation, parameter validation, special token stripping
Autoresearch deep testing (20 rounds across OpenAI SDK, Aider, LangChain,
LiteLLM, Cline, OpenCode patterns) found 8 bugs:
- P1: stop sequences not forwarded to stream_generate()/generate() in SimpleEngine
- P1: stop sequences included in output (OpenAI spec requires truncation)
- P2: n > 1 silently ignored — now returns 400
- P2: negative max_tokens accepted — now returns 400
- P2: temperature out of 0-2 range accepted — now returns 400
- P2: <|eom_id|>/<|python_tag|> Llama tokens leaking into responses
- P2: uvicorn keep-alive too short for agentic long-poll clients (now 30s)
- P3: completion_tokens off-by-one in non-streaming (re-encoding vs actual)
All 8 fixes verified against live server. 195/197 unit tests pass
(2 require live server on port 8000).
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
0 commit comments