Visible to the public Audio-CAPTCHA with distinction between random phoneme sequences and words spoken by multi-speaker

TitleAudio-CAPTCHA with distinction between random phoneme sequences and words spoken by multi-speaker
Publication TypeConference Paper
Year of Publication2017
AuthorsYamaguchi, M., Kikuchi, H.
Conference Name2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC)
KeywordsCAPTCHA, captchas, composability, distortion, Human Behavior, human beings, human factors, Proposals, pubcrawl, Semantics, Speech, Speech recognition, usability
AbstractAudio-CAPTCHA prevents malicious bots from attacking Web services and provides Web accessibility for visually-impaired persons. Most of the conventional methods employ statistical noise to distort sounds and let users remember and spell the words, which are difficult and laborious work for humans. In this paper, we utilize the difficulty on speaker-independent recognition for ASR machines instead of distortion with statistical noise. Our scheme synthesizes various voices by changing voice speed, pitch and native language of speakers. Moreover, we employ semantic identification problems between random phoneme sequences and meaningful words to release users from remembering and spelling words, so it improves the accuracy of humans and usability. We also evaluated our scheme in several experiments.
DOI10.1109/SMC.2017.8123098
Citation Keyyamaguchi_audio-captcha_2017