Spoken language-based automatic cognitive assessment of stroke survivors
Bahman Mirheidari , Simon M. Bell , Kirsty Harkness , Daniel Blackburn , Heidi Christensen
Language and Health ›› 2024, Vol. 2 ›› Issue (1) : 32 -38.
Stroke survivors (SSs) often experience cognitive decline following their initial stroke, necessitating repeat post-stroke cognitive assessments. Current methods of assessment, such as the pen-and-paper-based Montreal Cognitive Assessment (MoCA), is time-consuming and often reliant on seeing skilled clinicians in person. This is at a time when patients have a lot of often diverse rehabilitation needs. To address these challenges, our paper introduces the first system of its kind to be used for this cohort. CognoSpeak is an automated cognitive assessment system that people can use initially on the ward immediately post-stroke (baseline) and subsequently at home (follow-ups). CognoSpeak assesses cognitive decline by asking users to engage with a virtual agent by answering questions and completing clinically-motivated tasks and cognitive tests. The system then uses AI to extract and process speech, language, and interactional cues for cognitive decline. The system was originally developed for dementia; here, we show that it can successfully predict MoCA scores (regression) and identify cognitive decline predicated on a MoCA-based threshold (classification) in the stroke survivor cohort. We explore an extensive set of acoustic- and text-based features as well as different machine learning models. Leveraging a unique dataset of 55 SS CognoSpeak interactions, our findings show excellent performance for both regression and classification style prediction with the best regression result (Normalised Root Mean Squared Error (N-RMSE)) of 0.092. In addition, we show that direct classification of the MoCA score cutoff of 26 yields an F1-score of 0.74 (Specificity: 0.73, Sensitivity: 0.75) using a Logistic Regression Classifier. This demonstrates the first evidence of the system ’ s robustness and clinical potential.
Speech Technology / Post-stroke Rehabilitation / Cognitive decline assessment
| [1] |
|
| [2] |
Bandini, A., Green, J.R., Richburg, B., Yunusova, Y., (2018). Automatic detection of orofacial impairment in stroke.In: Interspeech, pp. 1711-1715. |
| [3] |
|
| [4] |
|
| [5] |
|
| [6] |
|
| [7] |
Devlin, J., Chang, M.-W., Lee, K., Toutanova, K., (2018). Bert: Pre-training of deep bidirectional transformers for language understanding, arXiv preprint arXiv: 1810.04805. |
| [8] |
|
| [9] |
|
| [10] |
Eyben, F., |
| [11] |
|
| [12] |
|
| [13] |
Fu, Z., Haider, F., Luz, S., (2020). Predicting mini-mental status examination scores through paralinguistic acoustic features of spontaneous speech, In:Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). IEEE, 5548-5552. |
| [14] |
|
| [15] |
|
| [16] |
|
| [17] |
|
| [18] |
|
| [19] |
|
| [20] |
|
| [21] |
Luz, S., Haider, F., |
| [22] |
Luz, S., Haider, F., |
| [23] |
|
| [24] |
Manohar, V., Povey, D., Khudanpur, S., (2017). JHU Kaldi system for Arabic MGB-3 ASR challenge using diarization, audio-transcript alignment and transfer learning, In:Proceedings of the 2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU). IEEE, pp. 346-352. |
| [25] |
Mirheidari, B., Blackburn, D., Christensen, H., (2022). Automatic cognitive assessment: Combining sparse datasets with disparate cognitive scores, in Proc. Interspeech. ISCA, 2022. |
| [26] |
|
| [27] |
|
| [28] |
Mirheidari, B., Blackburn, D., |
| [29] |
|
| [30] |
|
| [31] |
|
| [32] |
|
| [33] |
Pan, Y., Mirheidari, B., Harris, J.M., Thompson, J.C., Jones, M., Snowden, J.S., Blackburn, D., Christensen, H., (2021). Using the outputs of different automatic speech recognition paradigms for acoustic-and bert-based alzheimer ’ s dementia detection through spontaneous speech, Proc. Interspeech, pp. 3810-3814. |
| [34] |
|
| [35] |
|
| [36] |
|
| [37] |
|
| [38] |
|
| [39] |
Pennington, J., Socher, R., Manning, C., (2014). Glove:Global vectors for word representation, In: Proc. EMNLP, pp. 1532-1543. |
| [40] |
|
| [41] |
|
| [42] |
Romana, A., Bandon, J., Perez, M., Gutierrez, S., Richter, R., Roberts, A., Provost, E. M., (2021). Automatically detecting errors and disfluencies in read speech to predict cognitive impairment in people with Parkinson ’ s Disease, In:Proceedings of the INTERSPEECH 2021. International Speech Communication Association, pp. 156-160. |
| [43] |
|
| [44] |
|
| [45] |
|
| [46] |
Schuller, B., Steidl, S., Batliner, A., Hirschberg, J., Burgoon, J.K., Baird, A., Elkins, A., Zhang, Y., Coutinho, E., Evanini, K., (2016). The interspeech 2016 computational paralinguistics challenge: Deception, sincerity & native language, In:Proceedings of the 17TH Annual Conference of the International Speech Communication Association ( Interspeech 2016), Vols 1-5, vol. 8. ISCA, vol. 8. ISCA, 2001-2005. |
| [47] |
|
| [48] |
Sun, L., Zheng, J., Li, J., Qian, C., (2022). Exploring mmse score prediction model based on spontaneous speech.In: SEKE, 347-350. |
| [49] |
Triantafyllopoulos, A., Keren, G., Wagner, J., Steiner, I., Schuller, B., (2019). Towards robust speech emotion recognition using deep residual networks for speech enhancement. |
| [50] |
|
| [51] |
Valstar, M., Schuller, B., Smith, K., Eyben, F., Jiang, B., Bilakhia, S., Schnieder, S., Cowie, R., Pantic, M., (2013). Avec 2013: the continuous audio/visual emotion and depression recognition challenge, In: Proceedings of the 3rd ACM international workshop on Audio/visual emotion challenge, 3-10. |
| [52] |
Yancheva, M., Fraser, K.C., Rudzicz, F., (2015). Using linguistic features longitudinally to predict clinical scores for alzheimer ’ s disease and related dementias, In: Proceedings of SLPAT 2015: 6th Workshop on Speech and Language Processing for Assistive Technologies, 134-139. |
the Rosetrees Trust and the Stoneygate Trust (COMPASS, Grant Agreement No. M934). An NIHR Academic Clinical Lectureship in Neurology CL-2020-04-004 NIHR supports SMB. This summarises independent research at the NIHR Sheffield Biomedical Research Centre(Translational Neuroscience)
/
| 〈 |
|
〉 |