TR2018-113

VLASE: Vehicle Localization by Aggregating Semantic Edges


    •  Yu, X., Chaturvedi, S., Feng, C., Taguchi, Y., Lee, T.-Y., Fernandes, C., Ramalingam, S., "VLASE: Vehicle Localization by Aggregating Semantic Edges", IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), DOI: 10.1109/​IROS.2018.8594358, October 2018, pp. 3196-3203.
      BibTeX TR2018-113 PDF
      • @inproceedings{Yu2018oct,
      • author = {Yu, Xin and Chaturvedi, Sagar and Feng, Chen and Taguchi, Yuichi and Lee, Teng-Yok and Fernandes, Clinton and Ramalingam, Srikumar},
      • title = {VLASE: Vehicle Localization by Aggregating Semantic Edges},
      • booktitle = {IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)},
      • year = 2018,
      • pages = {3196--3203},
      • month = oct,
      • doi = {10.1109/IROS.2018.8594358},
      • url = {https://www.merl.com/publications/TR2018-113}
      • }
  • Research Areas:

    Artificial Intelligence, Computer Vision, Machine Learning

Abstract:

We propose VLASE, a framework to use semantic edge features from images to achieve on-road localization. Semantic edge features denote edge contours that separate pairs of distinct objects such as building-sky, road-sidewalk, and building-ground. While prior work has shown promising results by utilizing the boundary between prominent classes such as sky and building using skylines, we generalize this to consider 19 semantic classes. We extract semantic edge features using CASENet architecture and utilize VLAD framework to perform image retrieval. We achieve improvement over state of-the-art localization algorithms such as SIFT-VLAD and its deep variant NetVLAD. Ablation study shows the importance of different semantic classes, and our unified approach achieves better performance compared to individual prominent features such as skylines. We also introduce SLC Marathon dataset, a challenging dataset covering most of Salt Lake City with sufficient lighting variations.