We propose a system for user-aided visual localization of desert imagery without the use of any metadata such as GPS readings, camera focal length, or field-of-view. The system makes use only of publicly available datasets—in particular, digital elevation models (DEMs)—to rapidly and accurately locate photographs in nonurban environments such as deserts. Our system generates synthetic skyline views from a DEM and extracts stable concavity-based features from these skylines to form a database. To localize queries, a user manually traces the skyline on an input photograph. The skyline is automatically refined based on this estimate, and the same concavity-based features are extracted. We then apply a variety of geometrically constrained matching techniques to efficiently and accurately match the query skyline to a database skyline, thereby localizing the query image. We evaluate our system using a test set of 44 ground-truthed images over a 10,000 km2 region of interest in a desert and show that in many cases, queries can be localized with precision as fine as 100 m2 .
CITATION STYLE
Tzeng, E., Zhai, A., Clements, M., Townshend, R., & Zakhor, A. (2016). User-Aided Geo-location of Untagged Desert Imagery. In Advances in Computer Vision and Pattern Recognition (pp. 239–254). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-319-25781-5_13
Mendeley helps you to discover research relevant for your work.