Skip to content

[Feature] metrics support #3534

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 10 commits into
base: main
Choose a base branch
from
Draft

Conversation

CUHKSZzxy
Copy link
Collaborator

@CUHKSZzxy CUHKSZzxy commented May 9, 2025

Objective

Align with vLLM v1 metrics system and beyond. We also refer to SGLang monitoring.

TODO

  • Change time.perf_counter()
  • Change API server request arrival timestamp position
  • Abstract output processing outside of async engine generate()
  • Expert information collections
  • Grafana visualization

Usage

Start the server with --enable-metrics

lmdeploy serve api_server models--Qwen--Qwen2.5-7B-Instruct --enable-metrics
  • Metrics Publishing - Logging
    Information will be printed on the terminal every 10 seconds

  • Metrics Publishing - Prometheus & Grafana (WIP)
    Open http://xxxx:23333/metrics/ to view Prometheus details.

Related Issues & PR

Issue 2638, Issue 2673, PR1423

CUHKSZzxy added 2 commits May 9, 2025 20:38
Conflicts:
	lmdeploy/messages.py
	lmdeploy/pytorch/engine/engine.py
	lmdeploy/pytorch/engine/engine_instance.py
	lmdeploy/pytorch/messages.py
	lmdeploy/pytorch/paging/scheduler.py
@CUHKSZzxy CUHKSZzxy added the WIP label May 9, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant