Logo Kérwá
 

Enhancing speech recorded from a wearable sensor using a collection of autoencoders

dc.creatorGonzález Salazar, Astryd
dc.creatorGutiérrez Muñoz, Michelle
dc.creatorCoto Jiménez, Marvin
dc.date.accessioned2022-03-23T21:33:01Z
dc.date.available2022-03-23T21:33:01Z
dc.date.issued2020
dc.descriptionPart of the Communications in Computer and Information Science book series (CCIS, volume 1087).es_ES
dc.description.abstractAssistive Technology (AT) is a concept which includes the use of technological devices to improve the learning process or the general capabilities of people with disabilities. One of the major tasks of the AT is the development of devices that offer alternative or augmentative communication capabilities. In this work, we implemented a simple AT device with a low-cost sensor for registering speech signals, in which the sound is perceived as low quality and corrupted. Thus, it is not suitable to integrate into speech recognition systems, automatic transcription or general recognition of vocal-tract sounds for people with disabilities. We propose the use of a group of artificial neural networks that improve different aspects of the signal. In the study of the speech enhancement, it is normal to focus on how to make improvements in specific conditions of the signal, such as background noise, reverberation, natural noises, among others. In this case, the conditions that degrade the sound are unknown. This uncertainty represents a bigger challenge for the enhancement of the speech, in a real-life application. The results show the capacity of the artificial neural networks to enhance the quality of the sound, under several objective evaluation measurements. Therefore, this proposal can become a way of treating these kinds of signals to improve robust speech recognition systems and increase the real possibilities for implementing low-cost AT devices.es_ES
dc.description.procedenceUCR::Vicerrectoría de Docencia::Ingeniería::Facultad de Ingeniería::Escuela de Ingeniería Eléctricaes_ES
dc.description.sponsorshipUniversidad de Costa Rica/[322-B9-105]/UCR/Costa Ricaes_ES
dc.description.sponsorshipUniversidad de Costa Rica/[ED-3416]/UCR/Costa Ricaes_ES
dc.identifier.citationhttps://link.springer.com/chapter/10.1007/978-3-030-41005-6_26es_ES
dc.identifier.codproyecto322-B9-105
dc.identifier.codproyectoED-3416
dc.identifier.doi10.1007/978-3-030-41005-6_26
dc.identifier.isbn978-3-030-41005-6
dc.identifier.urihttps://hdl.handle.net/10669/86272
dc.language.isoenges_ES
dc.sourceHigh Performance Computing (pp.383-397).Turrialba, Costa Rica: Springer, Chames_ES
dc.subjectArtificial neural networkses_ES
dc.subjectAssistive Technologyes_ES
dc.subjectLong short-term memory (LSTM)es_ES
dc.subjectSpeech enhancementes_ES
dc.titleEnhancing speech recorded from a wearable sensor using a collection of autoencoderses_ES
dc.typecomunicación de congresoes_ES

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Springer3.pdf
Size:
1.82 MB
Format:
Adobe Portable Document Format
Description:

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
3.5 KB
Format:
Item-specific license agreed upon to submission
Description: