I'm doing inference for semantic segmentation using pretrained (on cityscapes dataset) DeepLab models based on xception_65 and xception_71 architectures. I've observed that:
- xception_65 is better in creating segmentation masks compared to xception_71.
- xception_71 is significantly faster than xception_65.
As xception_71 has 71 layers, which is more than the number of layers in xception_65 (65 layers), shouldn't it has higher inference time OR am I wrong somewhere?
(the number of blocks in xception_65 are fewer compared to that in xception_71)
You can check the code to reproduce the result at colab:segmentation_deeplab.ipynb.
%%time
print('Model:', MODEL_NAME)
seg_map = MODEL.run(original_im)
xception_65
Model: xception65_cityscapes_trainfine
CPU times: user 1.08 s, sys: 815 ms, total: 1.89 s
Wall time: 1.71 s
xception_71
Model: xception71_cityscapes_trainfine
CPU times: user 146 ms, sys: 28 ms, total: 174 ms
Wall time: 649 ms