Skip to content

šŸ—’ Feature request: ā€œConversation-as-a-Serviceā€ FastAPI demoĀ #67

@Flemington8

Description

@Flemington8

What problem does this solve?
In scenarios where LLM training occurs on a dedicated machine, and AIOpsLab operates on a separate server, there’s a need for a standardized interface to manage interactions between these components. Directly coupling the training loop with specific inference engines like vLLM can lead to rigid architectures that are difficult to maintain and scale. By introducing a RESTful API layer, you decouple the training process from the inference engine, promoting modularity and flexibility.

Proposed change

  • Add services/server.py (ā‰ˆ150 LOC) - FastAPI endpoint /simulate
  • Docs: docs/conversation_service.md with curl + python client snippet
  • Update README quick-start: uvicorn services.service:app --reload

Why FastAPI? FastAPI is fast (<2 ms overhead), self-documenting, and already widely used in
cloud/k8s. It introduces no heavy infra dependencies (only fastapi, pydantic, uvicorn).

Alternatives

  • gRPC (heavier; harder to demo)
  • websocket stream (nice-to-have later)

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions