Some Grafana Instances Taking Longer to Initialize
This incident has been resolved.
This incident has been resolved.
This incident has been resolved.
We continue to observe a continued period of recovery. At this time, we are considering this issue resolved.
On 12th of March, between 13:30 UTC and 13:50 UTC we experienced some degraded performance when querying metrics on the AWS Germany cluster. The effects were reflected in unavailability to see metrics on dashboards and alert queries triggering errors during that period. The issue has been resolved.
All series are now caught up and the incident is resolved.
This incident has been resolved.
Many ingesters were evicted from nodes in cortex-prod-01 at once causing a read path outage. Once the ingesters were rescheduled the read path recovered. The errors lasted about 10 minutes.
Our internal queues were affected by a scheduled data migration. For around three hours asynchronous scheduled tasks were affected and their processing delayed, but none cancelled or lost. Test scheduled particularly may have been not run at their intended time during that period.
This incident has been resolved.
This incident has been resolved.