Unstructured Audio Classification for Environment Recognition
Abstract
My thesis aims to contribute towards building autonomous agents that are able to understand their surrounding environment through the use of both audio and visual information. To capture a more complete description of a scene, the fusion of audio and visual information can be advantageous in enhancing the system’s context awareness. The goal of this work is on the characterization of unstructured environmental sounds for understanding and predicting the context surrounding of an agent. Most research on audio recognition has focused primarily on speech and music. Less attention has been paid to the challenges and opportunities for using audio to characterize unstructured environments. Unlike speech and music, which have formantic structures and harmonic structures,
Cite
Text
Chu. "Unstructured Audio Classification for Environment Recognition." AAAI Conference on Artificial Intelligence, 2008.Markdown
[Chu. "Unstructured Audio Classification for Environment Recognition." AAAI Conference on Artificial Intelligence, 2008.](https://mlanthology.org/aaai/2008/chu2008aaai-unstructured/)BibTeX
@inproceedings{chu2008aaai-unstructured,
title = {{Unstructured Audio Classification for Environment Recognition}},
author = {Chu, Selina},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2008},
pages = {1845-1846},
url = {https://mlanthology.org/aaai/2008/chu2008aaai-unstructured/}
}