A Multimodal Approach to Mapping Soundscapes
Abstract
We explore the problem of mapping soundscapes, that is, predicting the types of sounds that are likely to be heard at a given geographic location. Using a novel dataset, which includes geo-tagged audio and overhead imagery, we develop an approach for constructing an aural atlas, which captures the geospatial distribution of soundscapes. We build on previous work relating sound to ground-level imagery but incorporate overhead imagery to overcome the limitations of sparsely distributed geo-tagged audio. In the end, all that we require to construct an aural atlas is overhead imagery of the region of interest. We show examples of aural atlases at multiple spatial scales, from block-level to country.
Cite
Text
Salem et al. "A Multimodal Approach to Mapping Soundscapes." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.Markdown
[Salem et al. "A Multimodal Approach to Mapping Soundscapes." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.](https://mlanthology.org/cvprw/2018/salem2018cvprw-multimodal/)BibTeX
@inproceedings{salem2018cvprw-multimodal,
title = {{A Multimodal Approach to Mapping Soundscapes}},
author = {Salem, Tawfiq and Zhai, Menghua and Workman, Scott and Jacobs, Nathan},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2018},
pages = {2524-2527},
url = {https://mlanthology.org/cvprw/2018/salem2018cvprw-multimodal/}
}