-
Notifications
You must be signed in to change notification settings - Fork 3
Add Mimir Alertmanager alerts #1472
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This PR is missing the changelog entry as well as tests for the alerts. On top of that, could you add a comment explaining that those are coming from upstream like so? Does the dashboard that is linked in the alert exists in our grafanas? Stopping here but for the mimir alerts we put them all into the mimir alerts using another rule group: prometheus-rules/helm/prometheus-rules/templates/platform/atlas/alerting-rules/mimir.rules.yml Line 209 in 61773b2
I would advocate that we do the same :) |
Added a changelog entry, will add unit tests
Added a comment with a link to upstream
The dashboard uid is valid and points to our Mimir Alertmanager dashboard
I though having those alerts in another files would separate things and avoid a too long file for mimir.rules, but I moved them in there. |
if ! isInArray "$opsrecipe" "${opsRecipes[@]}"; then | ||
# or is a valid URL starting with http | ||
if ! isInArray "$opsrecipe" "${opsRecipes[@]}" && [[ "$opsrecipe" != http* ]]; then |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
What do we think about this ?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why not
I'm fine with splitting the alerts into multiple files but I would rather we have them in a folder then and that we do it for all alerts :) |
Towards: giantswarm/roadmap#3752
Adding some Mimir Alertmanager alerts to detect failures.
This a bit of a random pick, but we need something to start with and evaluate if things are running smoothly.
Those alerts are taken from upstream, therefore using upstream ops recipes until we have better ones.