Darknet: Why FN is negetive number and recall>1?

Created on 24 Apr 2018  路  3Comments  路  Source: AlexeyAB/darknet

I use "detector map" to evaluate the pretrained yolov3.weights on a person dataset where only persons are annotated. The result is like this
2018-04-24 2 26 02
Other classes are all zeros, which is reasonable, but in the end
2018-04-24 2 27 57
FN is negetive number and recall>1, I guess some detections hit the same ground truth, and FN is calculated by unique_truth_count-TP.
I think it's better to record whether each GT is detected and define FN as the number of GTs that are not detected.

Bug fixed

Most helpful comment

FN is negetive number and recall>1, I guess some detections hit the same ground truth

There were multiple hit for GT, so TP/FP/FN/IoU could be calculated wrong. (with mAP everything was fine.)
I fixed it. Now I check was this GT already detected or not. In any cases should be unique_truth_count=TP+FN

What TP/FP/FN/IoU can you get now?

All 3 comments

FN is negetive number and recall>1, I guess some detections hit the same ground truth

There were multiple hit for GT, so TP/FP/FN/IoU could be calculated wrong. (with mAP everything was fine.)
I fixed it. Now I check was this GT already detected or not. In any cases should be unique_truth_count=TP+FN

What TP/FP/FN/IoU can you get now?

I was running into the same problem, it appears to have been fixed now

Yes, it goes well now, thanks!
I'm very glad to see the recall of benchmark weights decreases by nearly a half, and my new-trained weights appears to work a little :)

Was this page helpful?
3 / 5 - 1 ratings

Related issues

PROGRAMMINGENGINEER-NIKI picture PROGRAMMINGENGINEER-NIKI  路  3Comments

kebundsc picture kebundsc  路  3Comments

qianyunw picture qianyunw  路  3Comments

jetsonnano2020 picture jetsonnano2020  路  3Comments

yongcong1415 picture yongcong1415  路  3Comments