Hi,
in Sentry i can sort by First Seen, Frequency, etc. but i cannot sort by user.
I know, detecting unique users can be hard, but the numbers are there already, when i can sort by user it would give me at least some hint how many people are affected.
A bug that happened for 10k users once is in my view more important than a bug that a happened 10k times for two user :)
PS: keep up the great work, sentry is great!
What if we just implemented it like this:

++
And what is the status of?
The user count that's displayed in the stream view is derived from the GroupTagKey (sentry_grouptagkey table) record. As far as I can tell, the biggest blocker here is that there is no index that would satisfy this query. Most (maybe all?) of the other sort clauses are directly off indexed[1] columns on the sentry_groupedmessage table.
There are a couple of options:
(project_id, key, values_seen DESC) on the sentry_grouptagkey table.Regardless, we'd have to create some additional index on this table in either case which could potentially be expensive. I do think this is a potentially useful feature, but not sure offhand what the cost of creating this index would be on write operations or on disk utilization.
Interested to hear if there has been any progress with this. Sorting by users still is one prominent wish for us and would greatly improve our work with Sentry.
Indeed. My organization judges issue severity by number of users impacted, rather than total occurrences.
I don't really understand why is this complicated? Cleary, when I do an API call to issue the response, does have userCount field. Can you just include this to you Sentry CLI then?
Anyways, I just made a small script for me to use. Let me know if anyone want this happy to share it. I hope the sentry team can add this as their feature.

Here is the script that I've been using to get the details out that I need, sorted by users affected:
curl -H 'Authorization: Bearer $SENTRY_TOKEN' $SENTRY_DOMAIN/api/0/projects/$ORGANIZATION/$PROJECT/issues/ | \
jq -r '[.[] | {id: .id, title: .title, count: .count | tonumber, users: .userCount, firstSeen: .firstSeen, lastSeen: .lastSeen, culprit: .culprit, filename: .metadata.filename, link: .permalink}] | sort_by(-.users)' | \
in2csv -f json > /tmp/errors.csv
Relies on jq and csvkit. If you want the output similar to @chocnut, instead of redirecting to a file, pipe to csvlook -y 0
Can't believe this was raised in 2016 and not yet addressed.
When ordered by Frequency there are cases when the error rate per user is higher than some other issues. For example on the screenshot below the 3rd highest had 15k events from 326 users where as the 5th most frequent issue had 19k events from 16k users. I would think that most people would prioritize the 19k events from 16k users issue rather then the 15k events from 326 users issue.

Anyone who still interested in or looking for a tool. Please check my repo here
Most helpful comment
Indeed. My organization judges issue severity by number of users impacted, rather than total occurrences.