python object_detection/eval.py \
--logtostderr \
--checkpoint_dir=ssd_mobilenet_v1_coco_2017_11_17 \
--eval_dir=$eval_dir \
--pipeline_config_path=object_detection/samples/configs/ssd_mobilenet_v1_coco.config
First bug, I cannot even evaluate model ssd_mobilenet_v1_coco_2017_11_17 without adding "metrics_set: coco_detection_metrics" in eval_config{} in object_detection/samples/configs/ssd_mobilenet_v1_coco.config
More important, I got mAP: 26, not 21 as in the offical table. Also for ssd_mobilenet_v2_coco_2018_03_29 I got 25, not 22 as in the official table.
Evaluation was on: COCO val_2017 (tfRecords are created by provided script ./object_detection/dataset_tools/download_and_preprocess_mscoco.sh )
Link on the official table: https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/detection_model_zoo.md
here is my output:
INFO:tensorflow:depth of additional conv before box predictor: 0
INFO:tensorflow:depth of additional conv before box predictor: 0
INFO:tensorflow:depth of additional conv before box predictor: 0
INFO:tensorflow:depth of additional conv before box predictor: 0
INFO:tensorflow:depth of additional conv before box predictor: 0
INFO:tensorflow:depth of additional conv before box predictor: 0
INFO:tensorflow:Restoring parameters from tiris/ssd_mobilenet_v1_coco_2017_11_17/model.ckpt
INFO:tensorflow:Restoring parameters from tiris/ssd_mobilenet_v1_coco_2017_11_17/model.ckpt
creating index...
index created!
INFO:tensorflow:Loading and preparing annotation results...
INFO:tensorflow:Loading and preparing annotation results...
INFO:tensorflow:DONE (t=0.39s)
INFO:tensorflow:DONE (t=0.39s)
creating index...
index created!
Running per image evaluation...
Evaluate annotation type bbox
DONE (t=71.17s).
Accumulating evaluation results...
DONE (t=11.80s).
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.263
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.419
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.279
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.016
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.132
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.508
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.238
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.342
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.362
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.042
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.253
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.643
@dextroza would you mind posting your actual command to reproduce this?
@ahundt check my post above, the actual command is added.
@dextroza , Hi Dear,i've trainined ssd_mobile_coco_v1 on my own dataset successful , but i have problem with running in eval.py , i saw you get result with eval.py , please give your experiences .
and my other problem evaluation depend on map coco , i get some error about this protocol.
@zeynali What problem do you have? Could you be more specific? I've just run their eval.py with paths to ckpt, config and eval_dir and it did evaluation successfully.
@dextroza , Hi , i've trained the ssd_mobilev1_coco on my own dataset that have only one class the total dataset for training is 200k samples and for testing 35k , and i go through 450k step with size 608*608 and batch_size 20 , and my train mAP is 85 but test is 35 why ??????
@zeynali You should check your dataset, data augmentation or try with L2 or dropout regularization if is possible. Let me know if you have any news. Good luck!
@dextroza similar with your result:
creating index...
index created!
INFO:tensorflow:Loading and preparing annotation results...
INFO:tensorflow:Loading and preparing annotation results...
INFO:tensorflow:DONE (t=0.27s)
INFO:tensorflow:DONE (t=0.27s)
creating index...
index created!
Running per image evaluation...
Evaluate annotation type bbox
DONE (t=43.80s).
Accumulating evaluation results...
DONE (t=7.45s).
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.263
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.421
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.278
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.017
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.131
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.510
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.240
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.343
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.362
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.043
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.253
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.644
@jch1 do you have time to check this, please?
I heard that 21 is from coco2014, while with coco2017 your number is correct. Can you test with coco2014?
can someone explain why do we get 6 precision and recall values each time?
Hi There,
We are checking to see if you still need help on this, as this seems to be considerably old issue. Please update this issue with the latest information, code snippet to reproduce your issue and error you are seeing.
If we don't hear from you in the next 7 days, this issue will be closed automatically. If you don't need help on this issue any more, please consider closing this.
Most helpful comment
@dextroza similar with your result:
creating index...
index created!
INFO:tensorflow:Loading and preparing annotation results...
INFO:tensorflow:Loading and preparing annotation results...
INFO:tensorflow:DONE (t=0.27s)
INFO:tensorflow:DONE (t=0.27s)
creating index...
index created!
Running per image evaluation...
Evaluate annotation type bbox
DONE (t=43.80s).
Accumulating evaluation results...
DONE (t=7.45s).
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.263
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.421
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.278
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.017
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.131
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.510
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.240
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.343
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.362
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.043
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.253
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.644