Perception of Visible Speech: Influence of Spatial Quantization

Abstract
Visible speech reading was studied to determine which features are functional and to test several models of pattern recognition. Nine test syllables differing in their initial consonant were presented in intact form or under various levels of spatial quantization. Performance decreased in increasing quantization but remained relatively good at moderate levels of degradation. Different models were tested against the confusion matrices. Six features were identified as functional in distinguishing among the nine consonant – vowel syllables. These features were used as sources of information in a fuzzy-logical model of perception and an additive model. The fuzzy-logical model provided a significantly better description of the confusion matrices, showing that speech reading is analogous to other domains of pattern recognition such as face recognition and facial-affect perception.