RESUMEN
PURPOSE: To describe the effects of subthalamic nucleus deep brain stimulation (STN-DBS) on the speech of Spanish-speaking Parkinson's disease (PD) patients during the first year of treatment. METHODS: The speech measures (SMs): maximum phonation time, acoustic voice measures, speech rate, speech intelligibility measures, and oral diadochokinesis rates of nine Colombian idiopathic PD patients (four females and five males; age = 63 ± 7 years; years of PD = 10 ± 7 years; UPDRS-III = 57 ± 6; H&Y = 2 ± 0.3) were studied in OFF and ON medication states before and every three months during the first year after STN-DBS surgery. Praat software and healthy native listeners' ratings were used for speech analysis. Statistical analysis tried to find significant differences in the SMs during follow-up (Friedman test) and between medication states (Wilcoxon paired test). Also, a pre-surgery variation interval (PSVI) of reference for every participant and SM was calculated to make an individual analysis of post-surgery variation. RESULTS: Non-significative post-surgery or medication state-related differences in the SMs were found. Nevertheless, individually, based on PSVIs, the SMs exhibited: no variation, inconsistent or consistent variation during post-surgery follow-up in different combinations, depending on the medication state. CONCLUSION: As a group, participants did not have a shared post-surgery pattern of change in any SM. Instead, based on PSVIs, the SMs varied differently in every participant, which suggests that in Spanish-speaking PD patients, the effects of STN-DBS on speech during the first year of treatment could be highly variable.
Asunto(s)
Estimulación Encefálica Profunda , Enfermedad de Parkinson , Núcleo Subtalámico , Humanos , Enfermedad de Parkinson/terapia , Enfermedad de Parkinson/fisiopatología , Masculino , Femenino , Persona de Mediana Edad , Anciano , Inteligibilidad del Habla/fisiología , Lenguaje , Trastornos del Habla/etiología , Trastornos del Habla/terapia , Habla/fisiología , Medición de la Producción del Habla , Resultado del TratamientoRESUMEN
Speech emotion recognition is key to many fields, including human-computer interaction, healthcare, and intelligent assistance. While acoustic features extracted from human speech are essential for this task, not all of them contribute to emotion recognition effectively. Thus, reduced numbers of features are required within successful emotion recognition models. This work aimed to investigate whether splitting the features into two subsets based on their distribution and then applying commonly used feature reduction methods would impact accuracy. Filter reduction was employed using the Kruskal-Wallis test, followed by principal component analysis (PCA) and independent component analysis (ICA). A set of features was investigated to determine whether the indiscriminate use of parametric feature reduction techniques affects the accuracy of emotion recognition. For this investigation, data from three databases-Berlin EmoDB, SAVEE, and RAVDES-were organized into subsets according to their distribution in applying both PCA and ICA. The results showed a reduction from 6373 features to 170 for the Berlin EmoDB database with an accuracy of 84.3%; a final size of 130 features for SAVEE, with a corresponding accuracy of 75.4%; and 150 features for RAVDESS, with an accuracy of 59.9%.
Asunto(s)
Emociones , Análisis de Componente Principal , Habla , Humanos , Emociones/fisiología , Habla/fisiología , Bases de Datos Factuales , Algoritmos , Reconocimiento de Normas Patrones Automatizadas/métodosRESUMEN
Emotion recognition through speech is a technique employed in various scenarios of Human-Computer Interaction (HCI). Existing approaches have achieved significant results; however, limitations persist, with the quantity and diversity of data being more notable when deep learning techniques are used. The lack of a standard in feature selection leads to continuous development and experimentation. Choosing and designing the appropriate network architecture constitutes another challenge. This study addresses the challenge of recognizing emotions in the human voice using deep learning techniques, proposing a comprehensive approach, and developing preprocessing and feature selection stages while constructing a dataset called EmoDSc as a result of combining several available databases. The synergy between spectral features and spectrogram images is investigated. Independently, the weighted accuracy obtained using only spectral features was 89%, while using only spectrogram images, the weighted accuracy reached 90%. These results, although surpassing previous research, highlight the strengths and limitations when operating in isolation. Based on this exploration, a neural network architecture composed of a CNN1D, a CNN2D, and an MLP that fuses spectral features and spectogram images is proposed. The model, supported by the unified dataset EmoDSc, demonstrates a remarkable accuracy of 96%.
Asunto(s)
Aprendizaje Profundo , Emociones , Redes Neurales de la Computación , Humanos , Emociones/fisiología , Habla/fisiología , Bases de Datos Factuales , Algoritmos , Reconocimiento de Normas Patrones Automatizadas/métodosRESUMEN
OBJECTIVE: This study aimed to compare the influence of four different maxillary removable orthodontic retainers on speech. MATERIAL AND METHODS: Eligibility criteria for sample selection were: 20-40-year subjects with acceptable occlusion, native speakers of Portuguese. The volunteers (n=21) were divided in four groups randomized with a 1:1:1:1 allocation ratio. The four groups used, in random order, the four types of retainers full-time for 21 days each, with a washout period of 7-days. The removable maxillary retainers were: conventional wraparound, wraparound with an anterior hole, U-shaped wraparound, and thermoplastic retainer. Three volunteers were excluded. The final sample comprised 18 subjects (11 male; 7 female) with mean age of 27.08 years (SD=4.65). The speech evaluation was performed in vocal excerpts recordings made before, immediately after, and 21 days after the installation of each retainer, with auditory-perceptual and acoustic analysis of formant frequencies F1 and F2 of the vowels. Repeated measures ANOVA and Friedman with Tukey tests were used for statistical comparison. RESULTS: Speech changes increased immediately after conventional wraparound and thermoplastic retainer installation, and reduced after 21 days, but not to normal levels. However, this increase was statistically significant only for the wraparound with anterior hole and the thermoplastic retainer. Formant frequencies of vowels were altered at initial time, and the changes remained in conventional, U-shaped and thermoplastic appliances after three weeks. CONCLUSIONS: The thermoplastic retainer was more harmful to the speech than wraparound appliances. The conventional and U-shaped retainers interfered less in speech. The three-week period was not sufficient for speech adaptation.
Asunto(s)
Estudios Cruzados , Retenedores Ortodóncicos , Humanos , Femenino , Masculino , Adulto , Diseño de Aparato Ortodóncico , Adulto Joven , Habla/fisiologíaRESUMEN
Diagnostic tests for Parkinsonism based on speech samples have shown promising results. Although abnormal auditory feedback integration during speech production and impaired rhythmic organization of speech are known in Parkinsonism, these aspects have not been incorporated into diagnostic tests. This study aimed to identify Parkinsonism using a novel speech behavioral test that involved rhythmically repeating syllables under different auditory feedback conditions. The study included 30 individuals with Parkinson's disease (PD) and 30 healthy subjects. Participants were asked to rhythmically repeat the PA-TA-KA syllable sequence, both whispering and speaking aloud under various listening conditions. The results showed that individuals with PD had difficulties in whispering and articulating under altered auditory feedback conditions, exhibited delayed speech onset, and demonstrated inconsistent rhythmic structure across trials compared to controls. These parameters were then fed into a supervised machine-learning algorithm to differentiate between the two groups. The algorithm achieved an accuracy of 85.4%, a sensitivity of 86.5%, and a specificity of 84.3%. This pilot study highlights the potential of the proposed behavioral paradigm as an objective and accessible (both in cost and time) test for identifying individuals with Parkinson's disease.
Asunto(s)
Retroalimentación Sensorial , Enfermedad de Parkinson , Habla , Humanos , Femenino , Masculino , Anciano , Enfermedad de Parkinson/fisiopatología , Enfermedad de Parkinson/diagnóstico , Persona de Mediana Edad , Habla/fisiología , Retroalimentación Sensorial/fisiología , Proyectos Piloto , Trastornos Parkinsonianos/fisiopatología , Estudios de Casos y ControlesRESUMEN
Pauses in speech are indicators of cognitive effort during language production and have been examined to inform theories of lexical, grammatical and discourse processing in healthy speakers and individuals with aphasia (IWA). Studies of pauses have commonly focused on their location and duration in relation to grammatical properties such as word class or phrase complexity. However, recent studies of speech output in aphasia have revealed that utterances of IWA are characterised by stronger collocations, i.e., combinations of words that are often used together. We investigated the effects of collocation strength and lexical frequency on pause duration in comic strip narrations of IWA and non-brain-damaged (NBD) individuals with part of speech (PoS; content and function words) as covariate. Both groups showed a decrease in pause duration within more strongly collocated bigrams and before more frequent content words, with stronger effects in IWA. These results are consistent with frameworks which propose that strong collocations are more likely to be processed as holistic, perhaps even word-like, units. Usage-based approaches prove valuable in explaining patterns of preservation and impairment in aphasic language production.
Asunto(s)
Afasia , Habla , Humanos , Afasia/fisiopatología , Habla/fisiología , Masculino , Femenino , Persona de Mediana Edad , Anciano , Adulto , LenguajeRESUMEN
THIS ARTICLE USES WORDS OR LANGUAGE THAT IS CONSIDERED PROFANE, VULGAR, OR OFFENSIVE BY SOME READERS. Hate speech detection in online social networks is a multidimensional problem, dependent on language and cultural factors. Most supervised learning resources for this task, such as labeled datasets and Natural Language Processing (NLP) tools, have been specifically tailored for English. However, a large portion of web users around the world speak different languages, creating an important need for efficient multilingual hate speech detection approaches. In particular, such approaches should be able to leverage the limited cross-lingual resources currently existing in their learning process. The cross-lingual transfer in this task has been difficult to achieve successfully. Therefore, we propose a simple yet effective method to approach this problem. To our knowledge, ours is the first attempt to create a multilingual embedding model specific to this problem. We validate the effectiveness of our approach by performing an extensive comparative evaluation against several well-known general-purpose language models that, unlike ours, have been trained on massive amounts of data. We focus on a zero-shot cross-lingual evaluation scenario in which we classify hate speech in one language without having access to any labeled data. Despite its simplicity, our embeddings outperform more complex models for most experimental settings we tested. In addition, we provide further evidence of the effectiveness of our approach through an ad hoc qualitative exploratory analysis, which captures how hate speech is displayed in different languages. This analysis allows us to find new cross-lingual relations between words in the hate-speech domain. Overall, our findings indicate common patterns in how hate speech is expressed across languages and that our proposed model can capture such relationships significantly.
Asunto(s)
Multilingüismo , Procesamiento de Lenguaje Natural , Humanos , Habla/fisiología , Lenguaje , OdioRESUMEN
Speech can be defined as the human ability to communicate through a sequence of vocal sounds. Consequently, speech requires an emitter (the speaker) capable of generating the acoustic signal and a receiver (the listener) able to successfully decode the sounds produced by the emitter (i.e., the acoustic signal). Time plays a central role at both ends of this interaction. On the one hand, speech production requires precise and rapid coordination, typically within the order of milliseconds, of the upper vocal tract articulators (i.e., tongue, jaw, lips, and velum), their composite movements, and the activation of the vocal folds. On the other hand, the generated acoustic signal unfolds in time, carrying information at different timescales. This information must be parsed and integrated by the receiver for the correct transmission of meaning. This chapter describes the temporal patterns that characterize the speech signal and reviews research that explores the neural mechanisms underlying the generation of these patterns and the role they play in speech comprehension.
Asunto(s)
Habla , Humanos , Habla/fisiología , Percepción del Habla/fisiología , Acústica del Lenguaje , PeriodicidadRESUMEN
Semantic verbal fluency (SVF) impairment is present in several neurological disorders. Although activation in SVF-related areas has been reported, how these regions are connected and their functional roles in the network remain divergent. We assessed SVF static and dynamic functional connectivity (FC) and effective connectivity in healthy participants using functional magnetic resonance imaging. We observed activation in the inferior frontal (IFG), middle temporal (pMTG) and angular gyri (AG), anterior cingulate (AC), insular cortex, and regions of the superior, middle, and medial frontal gyri (SFG, MFG, MidFG). Our static FC analysis showed a highly interconnected task and resting state network. Increased connectivity of AC with the pMTG and AG was observed for the task. The dynamic FC analysis provided circuits with connections similarly modulated across time and regions related to category identification, language comprehension, word selection and recovery, word generation, inhibition of speaking, speech planning, and articulatory planning of orofacial movements. Finally, the effective connectivity analysis provided a network that best explained our data, starting at the AG and going to the pMTG, from which there was a division between the ventral and dorsal streams. The SFG and MFG regions were connected and modulated by the MidFG, while the inferior regions formed the ventral stream. Therefore, we successfully assessed the SVF network, exploring regions associated with the entire processing, from category identification to word generation. The methodological approach can be helpful for further investigation of the SVF network in neurological disorders.
Asunto(s)
Mapeo Encefálico , Encéfalo , Imagen por Resonancia Magnética , Vías Nerviosas , Semántica , Humanos , Masculino , Femenino , Imagen por Resonancia Magnética/métodos , Adulto , Mapeo Encefálico/métodos , Vías Nerviosas/fisiología , Vías Nerviosas/diagnóstico por imagen , Adulto Joven , Encéfalo/fisiología , Encéfalo/diagnóstico por imagen , Conducta Verbal/fisiología , Habla/fisiología , Red Nerviosa/fisiología , Red Nerviosa/diagnóstico por imagenRESUMEN
Stuttering, affecting approximately 1% of the global population, is a complex speech disorder significantly impacting individuals' quality of life. Prior studies using electromyography (EMG) to examine orofacial muscle activity in stuttering have presented mixed results, highlighting the variability in neuromuscular responses during stuttering episodes. Fifty-five participants with stuttering and 30 individuals without stuttering, aged between 18 and 40, participated in the study. EMG signals from five facial and cervical muscles were recorded during speech tasks and analyzed for mean amplitude and frequency activity in the 5-15 Hz range to identify significant differences. Upon analysis of the 5-15 Hz frequency range, a higher average amplitude was observed in the zygomaticus major muscle for participants while stuttering (p < 0.05). Additionally, when assessing the overall EMG signal amplitude, a higher average amplitude was observed in samples obtained from disfluencies in participants who did not stutter, particularly in the depressor anguli oris muscle (p < 0.05). Significant differences in muscle activity were observed between the two groups, particularly in the depressor anguli oris and zygomaticus major muscles. These results suggest that the underlying neuromuscular mechanisms of stuttering might involve subtle aspects of timing and coordination in muscle activation. Therefore, these findings may contribute to the field of biosensors by providing valuable perspectives on neuromuscular mechanisms and the relevance of electromyography in stuttering research. Further research in this area has the potential to advance the development of biosensor technology for language-related applications and therapeutic interventions in stuttering.
Asunto(s)
Electromiografía , Músculos Faciales , Habla , Tartamudeo , Humanos , Electromiografía/métodos , Masculino , Adulto , Femenino , Tartamudeo/fisiopatología , Habla/fisiología , Músculos Faciales/fisiología , Músculos Faciales/fisiopatología , Fenómenos Biomecánicos/fisiología , Adulto Joven , Adolescente , Contracción Muscular/fisiologíaRESUMEN
Objective. In recent years, electroencephalogram (EEG)-based brain-computer interfaces (BCIs) applied to inner speech classification have gathered attention for their potential to provide a communication channel for individuals with speech disabilities. However, existing methodologies for this task fall short in achieving acceptable accuracy for real-life implementation. This paper concentrated on exploring the possibility of using inter-trial coherence (ITC) as a feature extraction technique to enhance inner speech classification accuracy in EEG-based BCIs.Approach. To address the objective, this work presents a novel methodology that employs ITC for feature extraction within a complex Morlet time-frequency representation. The study involves a dataset comprising EEG recordings of four different words for ten subjects, with three recording sessions per subject. The extracted features are then classified using k-nearest-neighbors (kNNs) and support vector machine (SVM).Main results. The average classification accuracy achieved using the proposed methodology is 56.08% for kNN and 59.55% for SVM. These results demonstrate comparable or superior performance in comparison to previous works. The exploration of inter-trial phase coherence as a feature extraction technique proves promising for enhancing accuracy in inner speech classification within EEG-based BCIs.Significance. This study contributes to the advancement of EEG-based BCIs for inner speech classification by introducing a feature extraction methodology using ITC. The obtained results, on par or superior to previous works, highlight the potential significance of this approach in improving the accuracy of BCI systems. The exploration of this technique lays the groundwork for further research toward inner speech decoding.
Asunto(s)
Interfaces Cerebro-Computador , Electroencefalografía , Habla , Humanos , Electroencefalografía/métodos , Electroencefalografía/clasificación , Masculino , Habla/fisiología , Femenino , Adulto , Máquina de Vectores de Soporte , Adulto Joven , Reproducibilidad de los Resultados , AlgoritmosRESUMEN
PURPOSE: To seek evidence of validity and reliability for the Compressed Speech Test with Figures. METHODS: The study was subdivided into three stages: construct validation, criteria and reliability. All participants were aged between 6:00 and 8:11. For the construct, Compressed Speech with Figures and the gold standard Adapted Compressed Speech test were applied to children with typical phonological development. For criterion analysis, Compressed Speech with Figures was applied in two groups, with typical (G1) and atypical (G2) phonological development. Finally, the application protocols underwent analysis by two Speech Therapists, with experience in the area of Central Auditory Processing, seeking to obtain an inter-evaluator reliability analysis. RESULTS: The correlation test indicated an almost perfect construct (correlation 0.843 for the right ear and 0.823 for the left ear). In the criterion analysis, it was noticed that both groups presented satisfactory results (G1 = 99.6 to 100%; G2 = 96 to 96.5%). The reliability analysis demonstrated that the protocol is easy to analyze, as both professionals presented unanimous responses. CONCLUSION: It was possible to obtain evidence of validity and reliability for the Compressed Speech with Figures instrument. The construct analysis showed that the instrument measures the same variable as the gold standard test, with an almost perfect correlation. In the criterion analysis, both groups presented similar performance, demonstrating that the instrument does not seem to differentiate populations with and without mild phonological disorder. The inter-evaluator reliability analysis demonstrated that the protocol is easy to analyze and score.
OBJETIVO: Buscar evidências de validade e fidedignidade para o Teste de Fala Comprimida com Figuras. MÉTODO: O estudo foi subdividido em três etapas: validação de construto, critério e fidedignidade. Todos os participantes tinham idade entre 6:00 e 8:11. Para o construto, aplicou-se o Fala Comprimida com Figuras e o teste padrão ouro Fala Comprimida Adaptado em crianças com desenvolvimento fonológico típico. Para análise de critério, aplicou-se o Fala Comprimida com Figuras em dois grupos, com desenvolvimento fonológico típico (G1) e atípico (G2). Por fim, os protocolos de aplicação passaram pela análise de duas Fonoaudiólogas, com experiência na área do Processamento Auditivo Central, buscando obter uma análise de fidedignidade interavaliadores. RESULTADOS: O teste de correlação indicou um construto quase perfeito (Rho=0,843 para orelha direita e Rho=0,823 para orelha esquerda). Na análise de critério, percebeu-se que ambos os grupos apresentaram resultados satisfatórios (G1 = 99,6 a 100%; G2 = 96 a 96,5%). Já a análise de fidedignidade demonstrou que o protocolo é de fácil análise, pois ambos os profissionais apresentaram respostas unânimes. CONCLUSÃO: Foi possível obter evidências de validade e fidedignidade para o instrumento de Fala Comprimida com Figuras. A análise de construto evidenciou que o instrumento mede a mesma variável que o teste padrão outro, com correlação quase perfeita. Na análise de critério, ambos os grupos apresentaram desempenho semelhante, demonstrando que o instrumento não parece diferenciar populações com e sem transtorno fonológico leve. A análise de fidedignidade interavaliador demonstrou que o protocolo é de fácil análise e pontuação.
Asunto(s)
Trastorno Fonológico , Habla , Niño , Humanos , Habla/fisiología , Reproducibilidad de los Resultados , Medición de la Producción del Habla , FonéticaRESUMEN
When engaged in a conversation, one receives auditory information from the other's speech but also from their own speech. However, this information is processed differently by an effect called Speech-Induced Suppression. Here, we studied brain representation of acoustic properties of speech in natural unscripted dialogues, using electroencephalography (EEG) and high-quality speech recordings from both participants. Using encoding techniques, we were able to reproduce a broad range of previous findings on listening to another's speech, and achieving even better performances when predicting EEG signal in this complex scenario. Furthermore, we found no response when listening to oneself, using different acoustic features (spectrogram, envelope, etc.) and frequency bands, evidencing a strong effect of SIS. The present work shows that this mechanism is present, and even stronger, during natural dialogues. Moreover, the methodology presented here opens the possibility of a deeper understanding of the related mechanisms in a wider range of contexts.
Asunto(s)
Electroencefalografía , Habla , Humanos , Habla/fisiología , Estimulación Acústica/métodos , Electroencefalografía/métodos , Encéfalo , Mapeo Encefálico/métodosRESUMEN
INTRODUCTION: The population of children with slow emergence of language development varies widely, both in their initial profile and in their response to intervention. In this sense, there is a group of late talkers who continue to show persistent language difficulties, in some cases exhibiting signs compatible with verbal dyspraxia. METHOD: In this paper we present the different response to intervention of two profiles of late talkers. Specifically, the Target Word© program (Hanen Centre) was implemented, which is addressed to latetalking children and their families. It combines the technique of focused stimulation with guidance to parents on strategies that stimulate global language development. RESULTS: Much of the symptomatology shown in one case of poor progress coincides with retrospective descriptions of children subsequently diagnosed with dyspraxia and can be considered early indicators of the disorder: unintelligibility, reduced consonant inventory or difficulties in word repetition. DISCUSSION: The different response to intervention contributes to diagnostic decision making and the early implementation of specific strategies directed to improve speech learning skills by incorporating motor learning principles. The few studies of intervention in suspected verbal dyspraxia in early childhood offer promising results on a variety of speech assessment indicators, and provide practitioners with valuable information with which to support the intervention in this population.
Introducción: La población de niños que comienzan con lentitud el desarrollo del lenguaje varía ampliamente, tanto en su perfil inicial como en la respuesta a la intervención. En este sentido, existe un grupo de niños, denominados hablantes tardíos, que continúan mostrando dificultades persistentes en el lenguaje. Algunos de estos niños muestran signos compatibles con la dispraxia verbal, y que se ponen de manifiesto a lo largo de la intervención. Método: En este trabajo presentamos la diferente respuesta a la intervención de dos perfiles de hablante tardío. Concretamente, se aplicó el programa Target Word©, del centro Hanen, que conjuga la técnica de la estimulación focalizada con la orientación a padres sobre estrategias que promueven el desarrollo del lenguaje. Resultados: Gran parte de la sintomatología mostrada en uno de los dos casos, que experimentó un progreso pobre, coincide con las descripciones retrospectivas de niños posteriormente diagnosticados con dispraxia y pueden considerarse indicadores tempranos del trastorno: ininteligibilidad, inventario consonántico reducido o dificultades en la repetición de palabras. Discusión: La diferente respuesta a la intervención contribuye a la toma de decisiones diagnósticas y a la aplicación temprana de estrategias específicas para la mejora de las habilidades de aprendizaje del habla mediante la incorporación de los principios del aprendizaje motor. Los escasos estudios de intervención en casos de sospecha de dispraxia verbal en la infancia temprana ofrecen resultados prometedores en diversos indicadores de evaluación del habla, y proporcionan a los profesionales una información valiosa en la que fundamentar la intervención en esta población.
Asunto(s)
Apraxias , Trastornos del Desarrollo del Lenguaje , Niño , Humanos , Preescolar , Habla/fisiología , Estudios Retrospectivos , Desarrollo del Lenguaje , Trastornos del Desarrollo del Lenguaje/diagnóstico , Trastornos del Desarrollo del Lenguaje/terapia , Apraxias/diagnóstico , Apraxias/terapiaRESUMEN
INTRODUCTION: The auditory perception of voice and its production involve auditory feedback, kinesthetic cues and the feedforward system that produce different effects for the voice. The Lombard, Sidetone and Pitch-Shift-Reflex effects are the most studied. The mapping of scientific experiments on changes in auditory feedback for voice motor control makes it possible to examine the existing literature on the phenomenon and may contribute to voice training or therapies. PURPOSE: To map experiments and research results with manipulation of auditory feedback for voice motor control in adults. METHOD: Scope review following the Checklist Preferred Reporting Items for Systematic reviews and Meta-Analyses extension (PRISMA-ScR) to answer the question: "What are the investigation methods and main research findings on the manipulation of auditory feedback in voice self-monitoring of adults?". The search protocol was based on the Population, Concept, and Context (PCC) mnemonic strategy, in which the population is adult individuals, the concept is the manipulation of auditory feedback and the context is on motor voice control. Articles were searched in the databases: BVS/Virtual Health Library, MEDLINE/Medical Literature Analysis and Retrieval System online, COCHRANE, CINAHL/Cumulative Index to Nursing and Allied Health Literature, SCOPUS and WEB OF SCIENCE. RESULTS: 60 articles were found, 19 on the Lombard Effect, 25 on the Pitch-shift-reflex effect, 12 on the Sidetone effect and four on the Sidetone/Lombard effect. The studies are in agreement that the insertion of a noise that masks the auditory feedback causes an increase in the individual's speech intensity and that the amplification of the auditory feedback promotes the reduction of the sound pressure level in the voice production. A reflex response to the change in pitch is observed in the auditory feedback, however, with particular characteristics in each study. CONCLUSION: The material and method of the experiments are different, there are no standardizations in the tasks, the samples are varied and often reduced. The methodological diversity makes it difficult to generalize the results. The main findings of research on auditory feedback on voice motor control confirm that in the suppression of auditory feedback, the individual tends to increase the intensity of the voice. In auditory feedback amplification, the individual decreases the intensity and has greater control over the fundamental frequency, and in frequency manipulations, the individual tends to correct the manipulation. The few studies with dysphonic individuals show that they behave differently from non-dysphonic individuals.
INTRODUçÃO: A percepção auditiva da voz e sua produção envolvem o feedback auditivo, as pistas cinestésicas e o sistema de feedforward, os quais produzem efeitos distintos para a voz. Os efeitos Lombard, Sidetone e o Pitch-Shift-Reflex são os mais estudados. O mapeamento de experimentos científicos sobre as modificações do feedback auditivo para o controle motor da voz possibilita examinar a literatura existente sobre o fenômeno e pode contribuir para o treinamento ou terapias da voz. OBJETIVO: Mapear os experimentos e resultados das pesquisas com manipulação do feedback auditivo para o controle motor da voz de indivíduos adultos. MÉTODO: Revisão de escopo seguindo o Checklist Preferred Reporting Items for Systematic reviews and Meta-Analyses extension (PRISMA-ScR) para responder à pergunta: "Quais os métodos de investigação e principais achados das pesquisas sobre a manipulação do feedback auditivo no automonitoramento da voz de indivíduos adultos?". O protocolo de busca foi baseado na estratégia mnemônica População, Conceito e Contexto (PCC). A população são os indivíduos adultos; o conceito é a manipulação do feedback auditivo e o contexto é o controle motor da voz. Os artigos foram pesquisados nas bases de dados: BVS/ Biblioteca Virtual em Saúde, MEDLINE/Medical Literature Analysis and Retrieval Sistem on-line, COCHRANE, CINAHL/Cumulative Index to Nursing and Allied Health Literature, SCOPUS e WEB OF SCIENCE. RESULTADOS: Foram encontrados 60 artigos, sendo 19 da temática do Efeito Lombard, 25 do efeito Pitch-shift-reflex, 12 do efeito Sidetone e quatro sobre o efeito Sidetone/Lombard. Os estudos são concordantes que a inserção de um ruído que mascara o feedback auditivo provoca um aumento na intensidade de fala do indivíduo e que a amplificação do feedback auditivo promove a redução do nível de pressão sonora na produção da voz. Observa-se uma resposta reflexa à mudança de tom no feedback auditivo, porém, com características individuais em cada estudo. CONCLUSÃO: O material e método dos experimentos são distintos, não há padronizações nas tarefas, as amostras são variadas, muitas vezes reduzidas. A diversidade metodológica dificulta a generalização dos resultados. Os principais achados das pesquisas a respeito o feedback auditivo sobre o controle motor da voz confirmam que, na supressão do feedback auditivo, o indivíduo tende a aumentar a intensidade da voz. Na amplificação do feedback auditivo, o indivíduo diminui a intensidade e tem maior controle sobre a frequência fundamental e, nas manipulações da frequência, o indivíduo tende a corrigir a manipulação. Os poucos estudos com sujeitos disfônicos mostram que eles se comportam diferentemente dos não disfônicos.
Asunto(s)
Percepción de la Altura Tonal , Voz , Adulto , Humanos , Retroalimentación , Percepción de la Altura Tonal/fisiología , Voz/fisiología , Habla/fisiología , Percepción AuditivaRESUMEN
Este estudio analizó las rutinas y los hábitos de alimentación de niños portugueses de entre 4 y 18 meses, los métodos de introducción de alimentos más utilizados y las percepciones de los padres acerca de la importancia de la introducción de alimentos en el desarrollo del habla, Se aplicó un cuestionario online a una muestra de 297 padres. El cuestionario mostró una alta fiabilidad (α = 0.86). La mayoría de las madres portuguesas respondió que amamantan a sus hijos o que lo hacen en combinación con mamadera. Además, que les presentan nuevos alimentos a sus hijos en forma de sopas o purés y que varían los sabores, texturas y consistencias con frecuencia. Aunque hubo resultados razonables con respecto al conocimiento sobre los métodos tradicionales y Baby Led Weaning (BLW), en esta muestra se observó una mayor frecuencia de uso del método tradicional en comparación con BLW. Por otro lado, no hubo una correlación significativa entre la edad de los padres y su percepción sobre la importancia de la alimentación para el habla. Sin embargo, los padres con estudios superiores mostraron una mayor percepción sobre la importancia de la alimentación en el habla. Los padres con un mayor número de hijos mostraron una menor consciencia sobre la importancia de la relación entre la alimentación y el habla.
This study analyzedthe eating routines and habits of Portuguese children aged 4 to 18 months, the methods of food introduction most frequently used, and the parents' perceptions regarding the importance of food introduction for speech development. A sample of 297 parents filled out an online self-report questionnaire about their perceptions, knowledge, and routines regarding food introduction methods, and their relationship with speech development. This questionnaire showed high reliability (α=0.86). The majority of the Portuguese mothers that filled out the survey answered either that they breastfed their children or breastfed them in combination with bottle feeding. They also stated that they introduce new food to their children in the form of soups or purees and that they vary the flavors, textures, and consistencies frequently. Although the results showed a reasonable level of knowledge regarding the traditional and Baby Led Weaning (BLW) methods, there was a higher frequency of use of the traditional method in this samplewhen compared to BLW. There was not a significant correlation between the age of the parents and their perception of the importance of feeding for speech. In turn, parents with a higher level of education showed a greater awareness of the relevance of feeding in the development of speech. Parents with a higher number of children showed less awareness of the relationship between feeding and speech development.
Asunto(s)
Humanos , Masculino , Femenino , Lactante , Adulto , Habla/fisiología , Métodos de Alimentación/psicología , Relaciones Padres-Hijo , Padres/psicología , Percepción , Portugal , Conocimientos, Actitudes y Práctica en Salud , Encuestas y Cuestionarios , Conducta Alimentaria/psicología , Autoinforme , Fenómenos Fisiológicos Nutricionales del LactanteRESUMEN
Introdução: O ceceio é um tipo de transtorno dos sons da fala decorrente de alterações nas estruturas orofaciais. Os métodos perceptivo-auditivos de avaliação fonoaudiológica podem gerar dúvidas quanto à natureza do transtorno e, portanto, as avaliações instrumentais são recomendadas para obter um diagnóstico mais preciso e completo. A ultrassonografia dos movimentos de língua permite a visualização em tempo real do movimento da língua durante a fala podendo contribuir na fonoterapia como biofeedback visual ultrassonográfico (BVU). Objetivos: Descrever os gestos articulatórios das fricativas /s/, /z/, /Ê/ e /Æ·/ pré e pós-terapia fonoaudiológica com BVU em uma criança com ceceio anterior. Métodos: Foram avaliados diversos aspectos da fala de uma menina de oito anos com ceceio anterior, e coletadas imagens ultrassonográficas antes e após cinco sessões de terapia utilizando BVU. Foram comparadas as imagens da língua na produção dos sons /s, z, Ê, Æ·/ pré e pós a intervenção fonoaudiológica com BVU. Resultados: Antes do tratamento, a paciente anteriorizava sem elevar a ponta de língua em /s/ e /z/; e em /Ê/ e /Æ·/ também havia anteriorização de ponto articulatório, mas sem interposição. Após a fonoterapia, houve ajuste do ponto articulatório e aquisição da elevação de ponta de língua em /s/ e /z/ resultando em maior constrição de língua em /Ê/ e /Æ·/. Conclusão: A ultrassonografia mostrou-se importante para a caracterização e descrição do ceceio anterior, e o seu uso como BVU permitiu melhora expressiva na produção articulatória por proporcionar automonitoramento durante a fala, num curto período de atendimento.
Introduction: lisping is a type of speech disorder resulting from changes in orofacial structures. The auditory-perceptual methods of speech-language assessment can raise doubts as to the nature of the disorder and, therefore, instrumental assessments are recommended to obtain a more accurate diagnosis. Ultrasonography of tongue movements allows real-time visualization of tongue movement during speech and may contribute to speech therapy as visual biofeedback. Objectives: To describe the articulatory gestures of fricatives /s/, /z/, /Ê/ e /Æ·/ before and after speech therapy with visual ultrasound biofeedback in a child with anterior lisp. Methods: Several aspects of the speech of an eight-year-old girl with previous lisp were evaluated, and ultrasound images were collected before and after 05 therapy sessions using ultrasound as instrumental biofeedback. The images of the tongue in the production of sounds /s, z, Ê, Æ·/ were compared before and after the speech therapy intervention. Results: Before treatment, the patient anteriorized without raising the tip of the tongue in /s/ and /z/; in /Ê/ and /Æ·/ there was anteriorization of the articulation point, but without interposition. After speech therapy, he adjusted the articulation point and acquired tongue tip elevation in /s/ and /z/ and presented greater tongue constriction in /Ê/ and /Æ·/. Conclusion: Ultrasonography proved to be important for the diagnosis and description of the previous lisp, and its use as biofeedback allowed a significant improvement in articulatory production due to its self-monitoring during speech, in a short period of care.
Introducción: el ceceo es un tipo de trastorno del habla resultante de cambios en las estructuras orofaciales. Los métodos auditivo-perceptivos de evaluación del habla y el lenguaje pueden generar dudas sobre la naturaleza del trastorno y, por lo tanto, se recomiendan evaluaciones instrumentales para obtener un diagnóstico más preciso. La ecografía de los movimientos de la lengua permite la visualización en tiempo real del movimiento de la lengua durante el habla y puede contribuir a la terapia del habla como biorretroalimentación visual. Objetivos: Describir los gestos articulatorios involucrados en la producción del habla de un niño con ceceo previo y compararlos antes y después de la logopedia con biofeedback visual por ultrasonido. Métodos: Se evaluaron varios aspectos del habla de una niña de ocho años con ceceo previo, y se recolectaron imágenes de ultrasonido antes y después de 05 sesiones de terapia utilizando ultrasonido como biofeedback instrumental. Se compararon las imágenes de la lengua en la producción de sonidos /s, z, Ê, Æ·/ antes y después de la intervención logopédica. Resultados: antes del tratamiento, el paciente realizó una anteriorización sin levantar la punta de la lengua en /s/ y /z/; en /Ê/ y /Æ·/ hubo anteriorización del punto de articulación, pero sin interposición. Después de la logopedia, ajustó el punto de articulación y adquirió elevación de la punta de la lengua en /s/ y /z/ y presentó mayor constricción de la lengua en /Ê/ y /Æ·/. Conclusión: La ecografía demostró ser importante para el diagnóstico y descripción del ceceo previo, y su uso como biofeedback permitió una mejora significativa en la producción articulatoria debido a su autocontrol durante el habla, en un corto período de atención.
Asunto(s)
Humanos , Femenino , Niño , Lengua/diagnóstico por imagen , Ultrasonografía , Trastorno Fonológico , Habla/fisiología , LogopediaRESUMEN
Connected speech is an everyday activity. We aimed to investigate whether connected speech can differentiate oral narrative production between adults with Alzheimer's disease (AD; nâ=â24) and cognitively healthy older adults (nâ=â48). We used graph attributes analysis to represent connected speech. Participants produced oral narratives and performed semantic, episodic, and working memory tasks. AD patients produced less connected narratives than cognitively healthy older adults. Connectedness was associated with semantic memory in AD and with episodic memory in controls. Word-graphs connectedness represents a practical tool to assess cognitive impairment in AD patients.
Asunto(s)
Enfermedad de Alzheimer/psicología , Trastornos de la Memoria/psicología , Memoria Episódica , Memoria a Corto Plazo/fisiología , Semántica , Habla/fisiología , Anciano , Anciano de 80 o más Años , Enfermedad de Alzheimer/diagnóstico , Femenino , Humanos , Masculino , Trastornos de la Memoria/diagnóstico , Persona de Mediana Edad , NarraciónRESUMEN
The purpose of this study was to explore the speaker-discriminatory potential of vowel formant mean frequencies in comparisons of identical twin pairs and non-genetically related speakers. The influences of lexical stress and the vowels' acoustic distances on the discriminatory patterns of formant frequencies were also assessed. Acoustic extraction and analysis of the first four speech formants F1-F4 were carried out using spontaneous speech materials. The recordings comprise telephone conversations between identical twin pairs while being directly recorded through high-quality microphones. The subjects were 20 male adult speakers of Brazilian Portuguese (BP), aged between 19 and 35. As for comparisons, stressed and unstressed oral vowels of BP were segmented and transcribed manually in the Praat software. F1-F4 formant estimates were automatically extracted from the middle points of each labeled vowel. Formant values were represented in both Hertz and Bark. Comparisons within identical twin pairs using the Bark scale were performed to verify whether the measured differences would be potentially significant when following a psychoacoustic criterion. The results revealed consistent patterns regarding the comparison of low-frequency and high-frequency formants in twin pairs and non-genetically related speakers, with high-frequency formants displaying a greater speaker-discriminatory power compared to low-frequency formants. Among all formants, F4 seemed to display the highest discriminatory potential within identical twin pairs, followed by F3. As for non-genetically related speakers, both F3 and F4 displayed a similar high discriminatory potential. Regarding vowel quality, the central vowel /a/ was found to be the most speaker-discriminatory segment, followed by front vowels. Moreover, stressed vowels displayed a higher inter-speaker discrimination than unstressed vowels in both groups; however, the combination of stressed and unstressed vowels was found even more explanatory in terms of the observed differences. Although identical twins displayed a higher phonetic similarity, they were not found phonetically identical.