Method Details


Details for method 'ScaleNet'

 

Method overview

name ScaleNet
challenge pixel-level semantic labeling
details The scale difference in driving scenarios is one of the essential challenges in semantic scene segmentation. Close objects cover significantly more pixels than far objects. In this paper, we address this challenge with a scale invariant architecture. Within this architecture, we explicitly estimate the depth and adapt the pooling field size accordingly. Our model is compact and can be extended easily to other research domains. Finally, the accuracy of our approach is comparable to the state-of-the-art and superior for scale problems. We evaluate on the widely used automotive dataset Cityscapes as well as a self-recorded dataset.
publication ScaleNet: Scale Invariant Network for Semantic Segmentation in Urban Driving Scenes
Mohammad Dawud Ansari, Stephan Krarß, Oliver Wasenmüller and Didier Stricker
International Conference on Computer Vision Theory and Applications, Funchal, Portugal, 2018
project page / code
used Cityscapes data fine annotations, coarse annotations
used external data ImageNet
runtime n/a
subsampling no
submission date October, 2017
previous submissions

 

Average results

Metric Value
IoU Classes 75.103
iIoU Classes 53.116
IoU Categories 89.6143
iIoU Categories 76.8198

 

Class results

Class IoU iIoU
road 98.3225 -
sidewalk 84.8265 -
building 92.3526 -
wall 50.0573 -
fence 59.6157 -
pole 62.7505 -
traffic light 71.7929 -
traffic sign 76.782 -
vegetation 93.1694 -
terrain 71.3511 -
sky 94.6282 -
person 83.6365 65.6485
rider 65.1556 44.7323
car 95.058 88.5791
truck 56.0106 34.6018
bus 71.6438 47.1669
train 59.8857 42.3924
motorcycle 66.2894 43.2404
bicycle 73.6294 58.5668

 

Category results

Category IoU iIoU
flat 98.4961 -
nature 92.8181 -
object 69.9455 -
sky 94.6282 -
construction 92.8909 -
human 84.1154 66.9017
vehicle 94.4057 86.738

 

Links

Download results as .csv file

Benchmark page