Un-actionable Issues in Sentry #41089
Replies: 10 comments 12 replies
-
Love that breakdown. I had a customer bring up a subtype of 2. They were unsure what was the origin of those and whether they should ignore them: 2.1 Client-side Mobile Connection errors.
The best suggestion I could come up at the time with was:
|
Beta Was this translation helpful? Give feedback.
-
Deployment related issues are a mixed bag. Sometimes things will spike/fail but only briefly during a deploy. Generally unhelpful as they are expected. |
Beta Was this translation helpful? Give feedback.
-
Excellent breakdown. We especially have issues with high volumes of
Most of these are some form of connection error - either:
We mentally classify these errors by downstream impact, mostly:
One other type of unactionable issue we have (sort of a subtype) is high volume un-actionable issues. These errors are expected to happen in high volumes, and as such we have to pre-batch them before sending to Sentry to not blow up our quota usage. For example, we have a caching service that may fail to read or write keys. This isn't a hard error (it's just a cache, who cares), unless the volumes get really high. Currently, if there is a temporary issue, we might generate a synthetic error to send to sentry, like CacheError("Cache key failure: failed to read 627 keys, write 329 keys, in the last minute") We did this because managing single error instances through the UI (and our quota) was untenable. If there was some hook we could plug into when initializing the sentry SDK that allowed us to id/batch error cases together, that would be a cool feature - or some other tooling to make bulk error spikes in this case feasible. As it happens now we've written our own layer, but it would be cool to see natively. |
Beta Was this translation helpful? Give feedback.
-
Excellent breakdown, which help me to deep think about issues in Sentry.
If a lagre number of evnets of this type are generated by the same user, it's very likely to be a script kiddies, so this type of issue is non-sensical. However, if this type of error is triggered by many people, even if only once per user, then i think it should be created by Sentry because it has a large impact.
|
Beta Was this translation helpful? Give feedback.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
This comment was marked as off-topic.
-
[Web only][may be off-topic] Sometimes what is essentially the same error get split into 3 issues because of legitimately different stacktraces (or lack thereof) between browsers: Technically they are duplicates, but:
|
Beta Was this translation helpful? Give feedback.
-
This is great.
|
Beta Was this translation helpful? Give feedback.
-
I'm also interested in this topic. Quite often a large service with many external integrations will produce errors you can not act on. We have sometimes removed them altogether from Sentry, but better would be to have them there, but as non-actionable issue. I'm wondering if categorising these to actionable and non-actionable issues would be the right way to go. TypeError would be an actionable issue. TimedOutError wouldn't typically be one. So, what should happen when an TimedOutError is first seen by Sentry? You would check it in UI, and click "not an actionable issue", and perhaps you would be asked to configure "raise as issue in case more than N in 5 minutes / AI autodetects spikes". So, now if you get an additional TimedOutError, it would still be recorded, but Sentry wouldn't show it as actionable. If you get a lot of TimedOutErrors, then Sentry would switch it to an actionable issue. After you systematically tackle all incoming issues in the UI by checking if they are actionable you are left with a list of actionable issues you should fix. Anything new you would check immediately to determine if it is trash (non-actionable), operational event (non-actionable with trigger on spike), or actionable. The goal for the team would be that you squash all actionable Sentry issues. |
Beta Was this translation helpful? Give feedback.
-
Sentry groups your errors into Issues so you find actual problems in your app. Issues offer a deeper understanding of the problem with error metadata such as stack traces, breadcrumbs, tags, etc.
However, some issues are ones you don’t care about and are often just an annoyance. We’re trying to characterize such Issues and use this to improve your Sentry experience by pruning low-quality issues from your Issue Stream and notifications.
We could categorize Sentry Issues into one of three buckets:
Help Sentry identify low-quality issues
We’d love to hear your thoughts about the third category, Un-actionable Issues. Some initial questions to help us on our way:
We’re sure this isn’t an exhaustive list of questions so definitely tell us anything else we’re missing!
Beta Was this translation helpful? Give feedback.
All reactions