Processing speech signal using auditory-like filterbank provides least uncertainty about articulatory gestures.
J Acoust Soc Am
; 129(6): 4014-22, 2011 Jun.
Article
en En
| MEDLINE
| ID: mdl-21682422
Understanding how the human speech production system is related to the human auditory system has been a perennial subject of inquiry. To investigate the production-perception link, in this paper, a computational analysis has been performed using the articulatory movement data obtained during speech production with concurrently recorded acoustic speech signals from multiple subjects in three different languages: English, Cantonese, and Georgian. The form of articulatory gestures during speech production varies across languages, and this variation is considered to be reflected in the articulatory position and kinematics. The auditory processing of the acoustic speech signal is modeled by a parametric representation of the cochlear filterbank which allows for realizing various candidate filterbank structures by changing the parameter value. Using mathematical communication theory, it is found that the uncertainty about the articulatory gestures in each language is maximally reduced when the acoustic speech signal is represented using the output of a filterbank similar to the empirically established cochlear filterbank in the human auditory system. Possible interpretations of this finding are discussed.
Texto completo:
1
Colección:
01-internacional
Base de datos:
MEDLINE
Asunto principal:
Vías Auditivas
/
Percepción del Habla
/
Medición de la Producción del Habla
/
Procesamiento de Señales Asistido por Computador
/
Cara
/
Gestos
/
Lenguaje
/
Modelos Teóricos
Tipo de estudio:
Diagnostic_studies
/
Prognostic_studies
Límite:
Female
/
Humans
/
Male
Idioma:
En
Revista:
J Acoust Soc Am
Año:
2011
Tipo del documento:
Article
País de afiliación:
Estados Unidos
Pais de publicación:
Estados Unidos