Phonotactic and acoustic cues for word segmentation in English
Phonotactic and acoustic cues for word segmentation in English. In Proceedings of the 6th International Conference on Spoken Language Processing (ICSLP 2000)
This study investigates the influence of both phonotactic and
acoustic cues on the segmentation of spoken English. Listeners
detected embedded English words in nonsense sequences (word
spotting). Words aligned with phonotactic boundaries were
easier to detect than words without such alignment. Acoustic
cues to boundaries could also have signaled word boundaries,
especially when word onsets lacked phonotactic alignment.
However, only one of several durational boundary cues showed
a marginally significant correlation with response times (RTs).
The results suggest that word segmentation in English is
influenced primarily by phonotactic constraints and only
secondarily by acoustic aspects of the speech signal.