MimirBucketIndexNotUpdated with multiple tenants #10729
Unanswered
TA3
asked this question in
Help and support
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have a self-hosted Mimir instances installed on a Kubernetes cluster with S3. Mimir is setup with mutil-tenancy enabled with around 12 tenants each pushing around 6000 samples per second.
For the past two months we've been seeing MimirBucketIndexNotUpdated alert, we've had a bunch of tweaking of the config based on some of the issues on Github, some of which has helped, but it's still happening at times. I did follow the steps provided in the runbook, that unfortunately didn't seem to help.
The only error logs that we see from the Mimir components are the following:
caller=retry.go:89 user=_local trace_id=4a50cf8a3ae421c7 level=error user=_local traceID=4a50cf8a3ae421c7 msg="error processing request" try=4 err="expanding series: the bucket index is too old. It was last updated at 2025-02-14T13:36:44Z, which exceeds the maximum allowed staleness period of 2h0m0s (err-mimir-bucket-index-too-old)"
caller=scheduler_processor.go:249 level=error user=_local traceID=4a50cf8a3ae421c7 msg="error notifying scheduler about finished query after the scheduler stream previously failed and returned error" err=EOF addr=10.244.76.246:9095 previousErr="rpc error: code = Canceled desc = context canceled: frontend cancelled query"
caller=bucket.go:450 level=error user=_local msg="loading block failed" elapsed=2.356089ms id=01JM0JHB7PJ658NBN75DBW5R7A err="create index header reader: write index header: new index reader: get object attributes of 01JM0JHB7PJ658NBN75DBW5R7A/index: The specified key does not exist."
From the compactor we get loads of these Warning messages again for pretty much all tenants:
caller=updater.go:131 level=warn component=cleaner run_id=1740402538 task=clean_up_users user=_local msg="skipped partial block when updating bucket index" block=01JHK9WTYWBZR00WKT8KB9QMWB
Mimir Config:
We have investigated our S3 (Minio) but there are no errors on there, Loki and Tempo also uses the same Minio instance with no issues at all.
Beta Was this translation helpful? Give feedback.
All reactions