Show simple item record

dc.contributor.authorSalomons, Inge
dc.contributor.authorDel Blanco Sierra, Eder
dc.contributor.authorNavas Cordón, Eva ORCID
dc.contributor.authorHernáez Rioja, Inmaculada ORCID
dc.contributor.authorDe Zuazo Oteiza, Xabier
dc.date.accessioned2023-08-08T12:15:09Z
dc.date.available2023-08-08T12:15:09Z
dc.date.issued2023-06-30
dc.identifier.citationApplied Sciences 13(13) : (2023) // Article ID 7746es_ES
dc.identifier.issn2076-3417
dc.identifier.urihttp://hdl.handle.net/10810/62135
dc.description.abstractThis paper evaluates the impact of inter-speaker and inter-session variability on the development of a silent speech interface (SSI) based on electromyographic (EMG) signals from the facial muscles. The final goal of the SSI is to provide a communication tool for Spanish-speaking laryngectomees by generating audible speech from voiceless articulation. However, before moving on to such a complex task, a simpler phone classification task in different modalities regarding speaker and session dependency is performed for this study. These experiments consist of processing the recorded utterances into phone-labeled segments and predicting the phonetic labels using only features obtained from the EMG signals. We evaluate and compare the performance of each model considering the classification accuracy. Results show that the models are able to predict the phonetic label best when they are trained and tested using data from the same session. The accuracy drops drastically when the model is tested with data from a different session, although it improves when more data are added to the training data. Similarly, when the same model is tested on a session from a different speaker, the accuracy decreases. This suggests that using larger amounts of data could help to reduce the impact of inter-session variability, but more research is required to understand if this approach would suffice to account for inter-speaker variability as well.es_ES
dc.description.sponsorshipThis research was funded by Agencia Estatal de Investigación grant number ref.PID2019-108040RB-C21/AEI/10.13039/501100011033es_ES
dc.language.isoenges_ES
dc.publisherMDPIes_ES
dc.relationinfo:eu-repo/grantAgreement/MICINN/PID2019-108040RB-C21es_ES
dc.rightsinfo:eu-repo/semantics/openAccesses_ES
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/
dc.subjectEMG signalses_ES
dc.subjectphone classificationes_ES
dc.subjectsilent speech interfaceses_ES
dc.subjecthuman–computer interactiones_ES
dc.subjectspeech processinges_ES
dc.titleFrame-Based Phone Classification Using EMG Signalses_ES
dc.typeinfo:eu-repo/semantics/articlees_ES
dc.date.updated2023-07-13T14:07:17Z
dc.rights.holder© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/ 4.0/).es_ES
dc.relation.publisherversionhttps://www.mdpi.com/2076-3417/13/13/7746es_ES
dc.identifier.doi10.3390/app13137746
dc.departamentoesIngeniería de comunicaciones
dc.departamentoeuKomunikazioen ingeniaritza


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/ 4.0/).
Except where otherwise noted, this item's license is described as © 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/ 4.0/).