Closed
Description
During ruler stop,it consume a lot memory, and probably killed by oom. In this case the information about itself is not unregister from the ring
below is the log captured by kubectl logs -f
level=info ts=2020-01-06T10:20:03.315150389Z caller=cortex.go:230 msg=initialising module=distributor
level=info ts=2020-01-06T10:20:03.315258575Z caller=cortex.go:230 msg=initialising module=ruler
level=info ts=2020-01-06T10:20:03.318200027Z caller=ruler.go:188 msg="ruler up and running"
level=info ts=2020-01-06T10:20:03.318295245Z caller=main.go:76 msg="Starting Cortex" version="(version=, branch=, revision=)"
level=info ts=2020-01-06T10:20:03.325922888Z caller=scheduler.go:246 msg="scheduler: updating rules for user" user_id=cidata num_groups=7 is_deleted=false
level=info ts=2020-01-06T10:20:03.326266046Z caller=lifecycler.go:460 msg="entry not found in ring, adding with no tokens"
level=info ts=2020-01-06T10:20:03.330674539Z caller=lifecycler.go:352 msg="auto-joining cluster after timeout"
level=error ts=2020-01-06T10:22:00.885438513Z caller=etcd.go:144 msg="error decoding key" key=ruler/ring err="snappy: corrupt input"
level=info ts=2020-01-06T10:22:03.333375976Z caller=lifecycler.go:599 msg="found empty ring, inserting tokens"
level=info ts=2020-01-06T10:22:25.497223122Z caller=signals.go:54 msg="=== received SIGINT/SIGTERM ===\n*** exiting"
level=info ts=2020-01-06T10:22:25.499116956Z caller=cortex.go:256 msg=stopping module=ruler
level=info ts=2020-01-06T10:22:25.503165591Z caller=notifier.go:602 msg="Stopping notification manager..."
level=info ts=2020-01-06T10:22:25.504910468Z caller=ruler.go:202 msg="shutting down rules scheduler"
level=info ts=2020-01-06T10:22:25.505102585Z caller=ruler.go:205 msg="waiting for workers to finish"
rpc error: code = Unknown desc = Error: No such container: 8c98cca62de0a0c28ba696daa5360f2f90c649ce9de178db5abcd04553156c8a%
Has anyone encountered this problem?