Open
Description
Is your feature request related to a problem?
In the k8s environment, when the pod (behind ClusterIP Service) is scaled up, the new pod has almost no requests.
Describe the solution you'd like
When the k8s pod is scaled up, the connection number between the new and old pods remains relatively balanced
Describe alternatives you've considered
For example, nginx sets an upper limit on the number of requests and a time limit for each connection. If the limit is exceeded, the connection is disconnected, so that new pods of the service will also have the opportunity to receive new requests.
keepalive_requests
keepalive_time
Related component
Client
Additional context
When I was stress testing the backend service, after the k8s pod was scaled up, the CPU usage of the new and old pods differed greatly because the new pods had much fewer connections than the old pods.
Code of Conduct
- I agree to follow the aio-libs Code of Conduct