Abstract
Originalsprache | englisch |
---|---|
Titel | Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) |
Publikationsstatus | Veröffentlicht - 2017 |
Veranstaltung | 2017 IEEE Conference on Computer Vision and Pattern Recognition - Honolulu, USA / Vereinigte Staaten Dauer: 21 Jul 2017 → 26 Jul 2017 |
Konferenz
Konferenz | 2017 IEEE Conference on Computer Vision and Pattern Recognition |
---|---|
Kurztitel | CVPR 2017 |
Land | USA / Vereinigte Staaten |
Ort | Honolulu |
Zeitraum | 21/07/17 → 26/07/17 |
Fingerprint
Dies zitieren
Learning to Align Semantic Segmentation and 2.5D Maps for Geolocalization. / Armagan, Anil; Hirzer, Martin; Roth, Peter M.; Lepetit, Vincent.
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). 2017.Publikation: Beitrag in Buch/Bericht/Konferenzband › Beitrag in einem Konferenzband › Forschung › Begutachtung
}
TY - GEN
T1 - Learning to Align Semantic Segmentation and 2.5D Maps for Geolocalization
AU - Armagan, Anil
AU - Hirzer, Martin
AU - Roth, Peter M.
AU - Lepetit, Vincent
PY - 2017
Y1 - 2017
N2 - We present an efficient method for geolocalization in urban environments starting from a coarse estimate of the location provided by a GPS and using a simple untextured 2.5D model of the surrounding buildings. Our key contribution is a novel efficient and robust method to optimize the pose: We train a Deep Network to predict the best direction to improve a pose estimate, given a semantic segmentation of the input image and a rendering of the buildings from this estimate. We then iteratively apply this CNN until converging to a good pose. This approach avoids the use of reference images of the surroundings, which are difficult to acquire and match, while 2.5D models are broadly available. We can therefore apply it to places unseen during training.
AB - We present an efficient method for geolocalization in urban environments starting from a coarse estimate of the location provided by a GPS and using a simple untextured 2.5D model of the surrounding buildings. Our key contribution is a novel efficient and robust method to optimize the pose: We train a Deep Network to predict the best direction to improve a pose estimate, given a semantic segmentation of the input image and a rendering of the buildings from this estimate. We then iteratively apply this CNN until converging to a good pose. This approach avoids the use of reference images of the surroundings, which are difficult to acquire and match, while 2.5D models are broadly available. We can therefore apply it to places unseen during training.
M3 - Conference contribution
BT - Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
ER -