Breaking Audio CAPTCHAs
Abstract
CAP T C H A s are computer-generated tests that humans can pass but current computer systems cannot. CAP T C H A s provide a method for automatically distinguishing a human from a computer program, and therefore can protect Web services from abuse by so-called "bots." Most CAP T C H A s consist of distorted images, usually text, for which a user must provide some description. Unfortunately, visual CAP T C H A s limit access to the millions of visually impaired people using the Web. Audio CAP T C H A s were created to solve this accessibility issue; however, the security of audio CAP T C H A s was never formally tested. Some visual CAP T C H A s have been broken using machine learning techniques, and we propose using similar ideas to test the security of audio CAP T C H A s . Audio CAP T C H A s are generally composed of a set of words to be identified, layered on top of noise. We analyzed the security of current audio CAP T CH A s from popular Web sites by using AdaBoost, SVM, and k-NN, and achieved correct solutions for test samples with accuracy up to 71%. Such accuracy is enough to consider these CAPTCHAs broken. Training several different machine learning algorithms on different types of audio CAP T C H A s allowed us to analyze the strengths and weaknesses of the algorithms so that we could suggest a design for a more robust audio CAPTCHA.
Cite
Text
Tam et al. "Breaking Audio CAPTCHAs." Neural Information Processing Systems, 2008.Markdown
[Tam et al. "Breaking Audio CAPTCHAs." Neural Information Processing Systems, 2008.](https://mlanthology.org/neurips/2008/tam2008neurips-breaking/)BibTeX
@inproceedings{tam2008neurips-breaking,
title = {{Breaking Audio CAPTCHAs}},
author = {Tam, Jennifer and Simsa, Jiri and Hyde, Sean and Ahn, Luis V.},
booktitle = {Neural Information Processing Systems},
year = {2008},
pages = {1625-1632},
url = {https://mlanthology.org/neurips/2008/tam2008neurips-breaking/}
}