Purpose: This research aimed to automatically predict intelligible speaking rate for individuals with Amyotrophic Lateral Sclerosis (ALS) based on speech acoustic and articulatory samples. Method: Twelve participants with ALS and two normal subjects produced a total of 1831 phrases. NDI Wave system was used to collect tongue and lip movement and acoustic data synchronously. A machine learning algorithm (i.e. support vector machine) was used to predict intelligible speaking rate (speech intelligibility × speaking rate) from acoustic and articulatory features of the recorded samples. Result: Acoustic, lip movement, and tongue movement information separately, yielded a R 2 of 0.652, 0.660, and 0.678 and a Root Mean Squared Error (RMSE) of 41.096, 41.166, and 39.855 words per minute (WPM) between the predicted and actual values, respectively. Combining acoustic, lip and tongue information we obtained the highest R 2 (0.712) and the lowest RMSE (37.562 WPM). Conclusion: The results revealed that our proposed analyses predicted the intelligible speaking rate of the participant with reasonably high accuracy by extracting the acoustic and/or articulatory features from one short speech sample. With further development, the analyses may be well-suited for clinical applications that require automatic speech severity prediction.

Automatic prediction of intelligible speaking rate for individuals with ALS from speech acoustic and articulatory samples / Wang, J.; Kothalkar, P. V.; Kim, M.; Bandini, A.; Cao, B.; Yunusova, Y.; Campbell, T. F.; Heitzman, D.; Green, J. R.. - In: INTERNATIONAL JOURNAL OF SPEECH-LANGUAGE PATHOLOGY. - ISSN 1754-9515. - 20:6(2018), pp. 669-679. [10.1080/17549507.2018.1508499]

Automatic prediction of intelligible speaking rate for individuals with ALS from speech acoustic and articulatory samples

Bandini A.;
2018

Abstract

Purpose: This research aimed to automatically predict intelligible speaking rate for individuals with Amyotrophic Lateral Sclerosis (ALS) based on speech acoustic and articulatory samples. Method: Twelve participants with ALS and two normal subjects produced a total of 1831 phrases. NDI Wave system was used to collect tongue and lip movement and acoustic data synchronously. A machine learning algorithm (i.e. support vector machine) was used to predict intelligible speaking rate (speech intelligibility × speaking rate) from acoustic and articulatory features of the recorded samples. Result: Acoustic, lip movement, and tongue movement information separately, yielded a R 2 of 0.652, 0.660, and 0.678 and a Root Mean Squared Error (RMSE) of 41.096, 41.166, and 39.855 words per minute (WPM) between the predicted and actual values, respectively. Combining acoustic, lip and tongue information we obtained the highest R 2 (0.712) and the lowest RMSE (37.562 WPM). Conclusion: The results revealed that our proposed analyses predicted the intelligible speaking rate of the participant with reasonably high accuracy by extracting the acoustic and/or articulatory features from one short speech sample. With further development, the analyses may be well-suited for clinical applications that require automatic speech severity prediction.
2018
20
6
669
679
Automatic prediction of intelligible speaking rate for individuals with ALS from speech acoustic and articulatory samples / Wang, J.; Kothalkar, P. V.; Kim, M.; Bandini, A.; Cao, B.; Yunusova, Y.; Campbell, T. F.; Heitzman, D.; Green, J. R.. - In: INTERNATIONAL JOURNAL OF SPEECH-LANGUAGE PATHOLOGY. - ISSN 1754-9515. - 20:6(2018), pp. 669-679. [10.1080/17549507.2018.1508499]
Wang, J.; Kothalkar, P. V.; Kim, M.; Bandini, A.; Cao, B.; Yunusova, Y.; Campbell, T. F.; Heitzman, D.; Green, J. R.
File in questo prodotto:
File Dimensione Formato  
2018_Wang_IJSLP.pdf

Accesso riservato

Tipologia: VOR - Versione pubblicata dall'editore
Licenza: [IR] closed
Dimensione 1.28 MB
Formato Adobe PDF
1.28 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
nihms-1514926.pdf

Open access

Tipologia: AAM - Versione dell'autore revisionata e accettata per la pubblicazione
Licenza: [IR] creative-commons
Dimensione 651.8 kB
Formato Adobe PDF
651.8 kB Adobe PDF Visualizza/Apri
Pubblicazioni consigliate

Licenza Creative Commons
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11380/1401628
Citazioni
  • ???jsp.display-item.citation.pmc??? 17
  • Scopus 35
  • ???jsp.display-item.citation.isi??? 72
social impact