Perceptual measures of speech from individuals with Parkinson's disease and multiple sclerosis: intelligibility and beyond.
Ontology highlight
ABSTRACT: The primary purpose of this study was to compare percent correct word and sentence intelligibility scores for individuals with multiple sclerosis (MS) and Parkinson's disease (PD) with scaled estimates of speech severity obtained for a reading passage.Speech samples for 78 talkers were judged, including 30 speakers with MS, 16 speakers with PD, and 32 healthy control speakers. Fifty-two naive listeners performed forced-choice word identification, sentence transcription, or visual analog scaling of speech severity for the Grandfather Passage (Duffy, 2005). Three expert listeners also scaled speech severity for the Grandfather Passage.Percent correct word and sentence intelligibility scores did not cleanly differentiate speakers with MS, PD, or control speakers. In contrast, both naive and expert listener groups judged reading passages produced by speakers with MS and PD to be more severely impaired than reading passages produced by control talkers.Scaled estimates of speech severity appear to be sensitive to aspects of speech impairment in MS and PD not captured by word or sentence intelligibility scores. One implication is that scaled estimates of speech severity might prove useful for documenting speech changes related to disease progression or even treatment for individuals with MS and PD with minimal reduction in intelligibility.
Project description:The perceptual consequences of rate reduction, increased vocal intensity, and clear speech were studied in speakers with multiple sclerosis (MS), Parkinson's disease (PD), and healthy controls.Seventy-eight speakers read sentences in habitual, clear, loud, and slow conditions. Sentences were equated for peak amplitude and mixed with multitalker babble for presentation to listeners. Using a computerized visual analog scale, listeners judged intelligibility or speech severity as operationally defined in Sussman and Tjaden (2012).Loud and clear but not slow conditions improved intelligibility relative to the habitual condition. With the exception of the loud condition for the PD group, speech severity did not improve above habitual and was reduced relative to habitual in some instances. Intelligibility and speech severity were strongly related, but relationships for disordered speakers were weaker in clear and slow conditions versus habitual.Both clear and loud speech show promise for improving intelligibility and maintaining or improving speech severity in multitalker babble for speakers with mild dysarthria secondary to MS or PD, at least as these perceptual constructs were defined and measured in this study. Although scaled intelligibility and speech severity overlap, the metrics further appear to have some separate value in documenting treatment-related speech changes.
Project description:Sensory and motor skills can be improved with training, but learning is often restricted to practice stimuli. As an exception, training on closed-loop (CL) sensorimotor interfaces, such as action video games and musical instruments, can impart a broad spectrum of perceptual benefits. Here we ask whether computerized CL auditory training can enhance speech understanding in levels of background noise that approximate a crowded restaurant. Elderly hearing-impaired subjects trained for 8 weeks on a CL game that, like a musical instrument, challenged them to monitor subtle deviations between predicted and actual auditory feedback as they moved their fingertip through a virtual soundscape. We performed our study as a randomized, double-blind, placebo-controlled trial by training other subjects in an auditory working-memory (WM) task. Subjects in both groups improved at their respective auditory tasks and reported comparable expectations for improved speech processing, thereby controlling for placebo effects. Whereas speech intelligibility was unchanged after WM training, subjects in the CL training group could correctly identify 25% more words in spoken sentences or digit sequences presented in high levels of background noise. Numerically, CL audiomotor training provided more than three times the benefit of our subjects' hearing aids for speech processing in noisy listening conditions. Gains in speech intelligibility could be predicted from gameplay accuracy and baseline inhibitory control. However, benefits did not persist in the absence of continuing practice. These studies employ stringent clinical standards to demonstrate that perceptual learning on a computerized audio game can transfer to "real-world" communication challenges.
Project description:Neural speech tracking has advanced our understanding of how our brains rapidly map an acoustic speech signal onto linguistic representations and ultimately meaning. It remains unclear, however, how speech intelligibility is related to the corresponding neural responses. Many studies addressing this question vary the level of intelligibility by manipulating the acoustic waveform, but this makes it difficult to cleanly disentangle effects of intelligibility from underlying acoustical confounds. Here, using magnetoencephalography (MEG) recordings, we study neural measures of speech intelligibility by manipulating intelligibility while keeping the acoustics strictly unchanged. Acoustically identical degraded speech stimuli (three-band noise vocoded, ~20 s duration) are presented twice, but the second presentation is preceded by the original (non-degraded) version of the speech. This intermediate priming, which generates a 'pop-out' percept, substantially improves the intelligibility of the second degraded speech passage. We investigate how intelligibility and acoustical structure affects acoustic and linguistic neural representations using multivariate Temporal Response Functions (mTRFs). As expected, behavioral results confirm that perceived speech clarity is improved by priming. TRF analysis reveals that auditory (speech envelope and envelope onset) neural representations are not affected by priming, but only by the acoustics of the stimuli (bottom-up driven). Critically, our findings suggest that segmentation of sounds into words emerges with better speech intelligibility, and most strongly at the later (~400 ms latency) word processing stage, in prefrontal cortex (PFC), in line with engagement of top-down mechanisms associated with priming. Taken together, our results show that word representations may provide some objective measures of speech comprehension.
Project description:A link between neurodegenerative diseases and infections has been previously reported. However, it is not clear to what extent such link is caused by confounding factors or to what extent it is intimately connected with the underlying conditions. Further, studies on the impact of infections on mortality risk following neurodegenerative diseases are rare. We analysed two data sets with different characteristics: (i) a community-based cohort from the UK Biobank with 2023 patients with multiple sclerosis, 2200 patients with Alzheimer's disease, 3050 patients with Parkinson's disease diagnosed before 1 March 2020 and 5 controls per case who were randomly selected and individually matched to the case; (ii) a Swedish Twin Registry cohort with 230 patients with multiple sclerosis, 885 patients with Alzheimer's disease and 626 patients with Parkinson's disease diagnosed before 31 December 2016 and their disease-free co-twins. The relative risk of infections after a diagnosis of neurodegenerative disease was estimated using stratified Cox models, with adjustment for differences in baseline characteristics. Causal mediation analyses of survival outcomes based on Cox models were performed to assess the impact of infections on mortality. Compared with matched controls or unaffected co-twins, we observed an elevated infection risk after diagnosis of neurodegenerative diseases, with a fully adjusted hazard ratio (95% confidence interval) of 2.45 (2.24-2.69) for multiple sclerosis, 5.06 (4.58-5.59) for Alzheimer's disease and 3.72 (3.44-4.01) for Parkinson's disease in the UK Biobank cohort, and 1.78 (1.21-2.62) for multiple sclerosis, 1.50 (1.19-1.88) for Alzheimer's disease and 2.30 (1.79-2.95) for Parkinson's disease in the twin cohort. Similar risk increases were observed when we analysed infections during the 5 years before diagnosis of the respective disease. Occurrence of infections after diagnosis had, however, relatively little impact on mortality, as mediation of infections on mortality (95% confidence interval) was estimated as 31.89% (26.83-37.11%) for multiple sclerosis, 13.38% (11.49-15.29%) for Alzheimer's disease and 18.85% (16.95-20.97%) for Parkinson's disease in the UK Biobank cohort, whereas it was 6.56% (-3.59 to 16.88%) for multiple sclerosis, -2.21% (-0.21 to 4.65%) for Parkinson's disease and -3.89% (-7.27 to -0.51%) for Alzheimer's disease in the twin cohort. Individuals with studied neurodegenerative diseases display an increased risk of infections independently of genetic and familial environment factors. A similar magnitude of risk increase is present prior to confirmed diagnosis, which may indicate a modulating effect of the studied neurological conditions on immune defences.
Project description:IntroductionAssessment of intelligibility in dysarthria tends to rely on oral reading of sentences or words. However, self-generated utterances are closer to a client's natural speech. This study investigated how transcription of utterances elicited by picture description can be used in the assessment of intelligibility in speakers with Parkinson's disease.MethodsSpeech samples from eleven speakers with Parkinson's disease and six neurologically healthy persons were audio-recorded. Forty-two naive listeners completed transcriptions of self-generated sentences from a picture description task and orally read sentences from the Swedish Test of Intelligibility, as well as scaled ratings of narrative speech samples.ResultsIntelligibility was higher in orally read than self-generated sentences and higher for content words than for the whole sentence in self-generated sentences for most of the speakers, although these within-group differences were not statistically significant at the group level. Adding contextual leads for the listeners increased intelligibility in self-generated utterances significantly but with individual variation. Although correlations between the intelligibility measures were at least moderate or strong, there was a considerable inter- and intra-speaker variability in intelligibility scores between tasks for the speakers with Parkinson's disease, indicating individual variation of factors that impact intelligibility. Intelligibility scores from neurologically healthy speakers were generally high across tasks with no significant differences between the conditions.ConclusionWithin-speaker variability supports literature recommendations to use multiple methods and tasks when assessing intelligibility. The inclusion of transcription of self-generated utterances elicited by picture description to the intelligibility assessment has the potential to provide additional information to assessment methods based on oral reading of pre-scripted sentences and to inform the planning of interventions.
Project description:BackgroundMore than 6,000,000 individuals worldwide are diagnosed with Parkinson's disease (PD). Nearly 90% develop speech signs that may substantially impair their speech intelligibility, resulting in losses in their communication and quality of life. Benefits of intensive speech treatment have been documented for a range of speech signs. However, the critical question of whether speech is more intelligible after treatment has not been investigated in a randomised controlled trial (RCT). We hypothesised that intensive speech treatment would improve speech intelligibility in PD.MethodSixty-four patients with hypokinetic dysarthria secondary to PD participated in this single-centre, parallel arm, statistically-powered RCT. Reporting follows CONSORT guidelines for non-pharmacological treatment. Patients were recruited from US clinics and randomised using a statistician-derived minimisation algorithm, to intensive speech treatment (16 1-hour sessions/1 month) targeting voice (voice group) or targeting articulation (articulation group) or to an untreated group (no treatment group). Speech treatments were delivered by speech clinicians who specialised in treating patients with PD. Trial design minimised bias and supported equipoise. For intelligibility assessment, blinded listeners (n = 117) orthographically transcribed 57 patients' recorded, self-generated narrative speech samples, randomly presented in multi-talker babble noise. Listeners were American-English speakers, ages 18-35 years, with normal hearing. The primary outcome was baseline (pre-treatment) to post-treatment change in transcription accuracy (TA), recognised as the most objective measure of intelligibility. TA was defined as the percentage of words transcribed correctly. Listeners, data collectors, and data managers were blinded to treatment conditions and groups. Reliability was evaluated using intraclass correlation coefficients and differences amongst groups were evaluated by mixed-effects models, in accordance with the intention-to-treat approach.This trial was registered with ClinicalTrials.gov Identifier: NCT00123084.FindingsBetween June 23, 2016 and August 14, 2017, blinded listeners transcribed baseline and post-treatment speech samples for intelligibility assessment of 57 patients in the voice (n = 19), articulation (n = 19) and no treatment (n = 19) groups. Between-group differences (d) in changes from baseline to post-treatment in TA indicated significantly greater increases following treatment targeting voice than treatment targeting articulation (d = 26·2%, 95% CI 1·5 - 51·0; p = 0·04; ES=1·0). Differences between TA changes in the treatment targeting voice and in the no treatment group were significant (d = 42·8%, 95% CI 22·4 - 63·2; p = 0·0002; ES=1·8). Differences between TA changes in the treatment targeting articulation and in the no treatment group were not significant (d = 16·5%, 95% CI -6·1 - 39·2; p = 0·147; ES=0·9).InterpretationThese findings provide the first RCT evidence that intensive speech treatment targeting voice improves speech intelligibility in PD. Thus, this evidence-based treatment may positively impact health-related quality of life for patients with PD globally when it is included in patient management.Funding
Project description:To investigate the effects of l-dopa (Levodopa) and cZi-DBS (deep brain stimulation in caudal zona incerta) on spontaneous speech intelligibility in patients with PD (Parkinson's disease).Spontaneous utterances were extracted from anechoic recordings from 11 patients with PD preoperatively (off and on l-dopa medication) and 6 and 12 months post bilateral cZi-DBS operation (off and on stimulation, with simultaneous l-dopa medication). Background noise with an amplitude corresponding to a clinical setting was added to the recordings. Intelligibility was assessed through a transcription task performed by 41 listeners in a randomized and blinded procedure.A group-level worsening in spontaneous speech intelligibility was observed on cZi stimulation compared to off 6 months postoperatively (8 adverse, 1 positive, 2 no change). Twelve months postoperatively, adverse effects of cZi-DBS were not frequently observed (2 positive, 3 adverse, 6 no change). l-dopa administered preoperatively as part of the evaluation for DBS operation provided the overall best treatment outcome (1 adverse, 4 positive, 6 no change).cZi-DBS was shown to have smaller negative effects when evaluated from spontaneous speech compared to speech effects reported previously. The previously reported reduction in word-level intelligibility 12 months postoperatively was not transferred to spontaneous speech for most patients. Reduced intelligibility due to cZi stimulation was much more prominent 6 months postoperatively than at 12 months.
Project description:Purpose Telemedicine, used to offset disparities in access to speech-language therapy, relies on technology that utilizes compression algorithms to transmit signals efficiently. These algorithms have been thoroughly evaluated on healthy speech; however, the effects of compression algorithms on the intelligibility of disordered speech have not been adequately explored. Method This case study assessed acoustic and perceptual effects of resampling and speech compression (i.e., transcoding) on the speech of 2 individuals with dysarthria. Forced-choice vowel identification and transcription tasks were utilized, completed by 20 naive undergraduate listeners. Results Results showed relative improvements and decrements in intelligibility, on various measures, based on the speakers' acoustic profiles. The transcoding of the speech compression algorithm resulted in an enlarged vowel space area and associated improvements in vowel identification for 1 speaker and a smaller vowel space area and decreased vowel identification for the other speaker. Interestingly, there was an overall decrease in intelligibility in the transcription task in this condition for both speakers. Conclusions There is a complex interplay between dysarthria and compression algorithms that warrants further exploration. The findings suggest that it is critical to be mindful of apparent changes in intelligibility secondary to compression algorithms necessary for practicing telemedicine. Supplemental Material https://doi.org/10.23641/asha.7291940.
Project description:PurposeTo determine the mechanisms of speech intelligibility impairment due to neurologic impairments, intelligibility decline was modeled as a function of co-occurring changes in the articulatory, resonatory, phonatory, and respiratory subsystems.MethodSixty-six individuals diagnosed with amyotrophic lateral sclerosis (ALS) were studied longitudinally. The disease-related changes in articulatory, resonatory, phonatory, and respiratory subsystems were quantified using multiple instrumental measures, which were subjected to a principal component analysis and mixed effects models to derive a set of speech subsystem predictors. A stepwise approach was used to select the best set of subsystem predictors to model the overall decline in intelligibility.ResultsIntelligibility was modeled as a function of five predictors that corresponded to velocities of lip and jaw movements (articulatory), number of syllable repetitions in the alternating motion rate task (articulatory), nasal airflow (resonatory), maximum fundamental frequency (phonatory), and speech pauses (respiratory). The model accounted for 95.6% of the variance in intelligibility, among which the articulatory predictors showed the most substantial independent contribution (57.7%).ConclusionArticulatory impairments characterized by reduced velocities of lip and jaw movements and resonatory impairments characterized by increased nasal airflow served as the subsystem predictors of the longitudinal decline of speech intelligibility in ALS. Declines in maximum performance tasks such as the alternating motion rate preceded declines in intelligibility, thus serving as early predictors of bulbar dysfunction. Following the rapid decline in speech intelligibility, a precipitous decline in maximum performance tasks subsequently occurred.
Project description:This study investigated the acoustic basis of within-speaker, across-utterance variation in sentence intelligibility for 12 speakers with dysarthria secondary to Parkinson's disease (PD). Acoustic measures were also obtained for 12 healthy controls for comparison to speakers with PD. Speakers read sentences using their typical speech style. Acoustic measures of speech rate, articulatory rate, fundamental frequency, sound pressure level and F2 interquartile range (F2 IQR) were obtained. A group of listeners judged sentence intelligibility using a computerized visual-analog scale. Relationships between judgments of intelligibility and acoustic measures were determined for individual speakers with PD. Relationships among acoustic measures were also quantified. Although considerable variability was noted, articulatory rate, fundamental frequency and F2 IQR were most frequently associated with within-speaker variation in sentence intelligibility. Results suggest that diversity among speakers with PD should be considered when interpreting results from group analyses.