Open
Description
Describe the bug
When executing a long backwards query (> query_ingesters_within
), we have the possibility of querying too much data from both the ingesters and the store, particularly if the ingesters' head block contains a large amount of unflushed logs which could satisfy the request.
To Reproduce
- have a simple, non-filtering query like
{cluster="us-central1"} | json | line_format "{{.msg}}"
- have a large number of unflushed lines in the head block (~7.5m in our case from all 130+ ingesters)
- execute the query over a 48h time range
- ingesters OOM
Expected behavior
The querier should query the ingesters for their unflushed data, and not need to examine chunks unnecessarily from the store.
Screenshots, Promtail config, or terminal output
Grafana Cloud trace ID: 410e61362f6ac7c2
Activity