chore(kubernetes): add podAntiAffinity#38954
Conversation
🧪 CI InsightsHere's what we observed from your CI run for 98f0c92. 🟢 All jobs passed!But CI Insights is watching 👀 |
|
Summary of ChangesHello @hongbo-miao, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request focuses on improving the deployment strategy for Grafana, NATS, and Valkey by adding topology spread constraints to their respective Helm charts. It also removes MLflow configurations. The topology spread constraints ensure that pods are distributed across the cluster to improve availability and resilience. Additionally, the replica count for Valkey has been reduced. Highlights
Changelog
Activity
Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Code Review
This pull request introduces pod anti-affinity and topology spread constraints to several services (Grafana, MLflow, NATS, Valkey) to improve their high availability by distributing pods across different nodes. This is a good practice. The PR also removes the MLflow application from the aws/production-hm environment. I've found one critical issue with a typo in a topology key that will likely cause deployment failure, and have also provided a recommendation regarding replica count for a stateful service. Please see the detailed comments.
...projects/nebius/production-hm/skypilot-valkey/environments/production/helm-chart/values.yaml
Show resolved
Hide resolved
📝 WalkthroughSummary by CodeRabbit
WalkthroughThe pull request removes the complete MLflow production deployment from AWS infrastructure and configures pod topology spread constraints and scheduling policies across multiple Kubernetes deployments in both AWS and Nebius environments for improved pod distribution. Changes
No actionable comments were generated in the recent review. 🎉 🧹 Recent nitpick comments
Tip Issue Planner is now in beta. Read the docs and try it out! Share your feedback on Discord. Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |



No description provided.