Logo Kérwá
 

Enhancing speech recorded from a wearable sensor using a collection of autoencoders

dc.creatorGonzález Salazar, Astryd
dc.creatorGutiérrez Muñoz, Michelle
dc.creatorCoto Jiménez, Marvin
dc.date.accessioned2022-03-23T21:33:01Z
dc.date.available2022-03-23T21:33:01Z
dc.date.issued2020
dc.descriptionPart of the Communications in Computer and Information Science book series (CCIS, volume 1087).es
dc.description.abstractAssistive Technology (AT) is a concept which includes the use of technological devices to improve the learning process or the general capabilities of people with disabilities. One of the major tasks of the AT is the development of devices that offer alternative or augmentative communication capabilities. In this work, we implemented a simple AT device with a low-cost sensor for registering speech signals, in which the sound is perceived as low quality and corrupted. Thus, it is not suitable to integrate into speech recognition systems, automatic transcription or general recognition of vocal-tract sounds for people with disabilities. We propose the use of a group of artificial neural networks that improve different aspects of the signal. In the study of the speech enhancement, it is normal to focus on how to make improvements in specific conditions of the signal, such as background noise, reverberation, natural noises, among others. In this case, the conditions that degrade the sound are unknown. This uncertainty represents a bigger challenge for the enhancement of the speech, in a real-life application. The results show the capacity of the artificial neural networks to enhance the quality of the sound, under several objective evaluation measurements. Therefore, this proposal can become a way of treating these kinds of signals to improve robust speech recognition systems and increase the real possibilities for implementing low-cost AT devices.es
dc.description.procedenceUCR::Vicerrectoría de Docencia::Ingeniería::Facultad de Ingeniería::Escuela de Ingeniería Eléctricaes
dc.description.sponsorshipUniversidad de Costa Rica/[322-B9-105]/UCR/Costa Ricaes
dc.description.sponsorshipUniversidad de Costa Rica/[ED-3416]/UCR/Costa Ricaes
dc.identifier.citationhttps://link.springer.com/chapter/10.1007/978-3-030-41005-6_26
dc.identifier.codproyecto322-B9105
dc.identifier.codproyectoED-3416
dc.identifier.doihttps://doi.org/10.1007/978-3-030-41005-6_26
dc.identifier.isbn978-3-030-41005-6
dc.identifier.urihttps://hdl.handle.net/10669/86272
dc.language.isoeng
dc.sourceHigh Performance Computing (pp.383-397).Turrialba, Costa Rica: Springer, Chames
dc.subjectArtificial neural networkses
dc.subjectAssistive Technologyes
dc.subjectLong short-term memory (LSTM)es
dc.subjectSpeech enhancementes
dc.titleEnhancing speech recorded from a wearable sensor using a collection of autoencoderses
dc.typecomunicación de congresoes

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Springer3.pdf
Size:
1.82 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
3.5 KB
Format:
Item-specific license agreed upon to submission
Description: