I ran a similar model with Titan Xp, however I used the infer_detections.py script and recorded the direct pass time [mainly using the date and time before and after tf_example = detect_inference.infer_detections_and_add_to_example (serialized_example_tensor, detected_boxes_tensor, detected_scores_tensestens_tensestens_tensestens_tensestens_tensestenselstenselstenslstensestensestensestenselstenselstensestensestensestenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenselstenslstenselstenselstenselstenselstenselstensestenselsensibletensorstsensors_tensorsts reduced the number of sentences generated in the first stage of FasterRCN from 300 to 100 and reduced the number of detections in the second stage to 100. I got numbers in the range from 80 to 140 ms, and I think that a 600x600 image will approximately take ~ 106 or a little less in this us the top three (due to the Titan Xp and the reduced complexity of the model). Perhaps you can repeat the above process on your equipment, so if in this case the numbers are also ~ 106 ms, we can explain the difference using the DockerFile and the client. If the numbers are still high, it might be hardware.
It would be helpful if someone from the Tensorflow Object Detection team could comment on the setting used to generate numbers in the model zoo .
source share