Publications

Detailed Information

Neural networks for compressing and classifying speaker-independent paralinguistic signals

Cited 3 time in Web of Science Cited 2 time in Scopus
Authors

Byun, Seokhyun; Yoon, Seunghyun; Jung, Kyomin

Issue Date
2019-04
Publisher
IEEE
Citation
2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP), pp.311-314
Abstract
Recognizing and classifying paralinguistic signals, with its various applications, is an important problem. In general, this task is considered challenging because the sound information from the signals is difficult to distinguish even by humans. Thus, analyzing signals with machine learning techniques is a reasonable approach to understanding signals. Audio features extracted from paralinguistic signals usually consist of high-dimensional vectors such as prosody, energy, cepstrum, and other speech-related information. Therefore, when the size of a training corpus is not sufficiently large, it is extremely difficult to apply machine learning methods to analyze these signals due to their high feature dimensions. This paper addresses these limitations by using neural networks' feature learning abilities. First, we use a neural network-based autoencoder to compress the signal to eliminate redundancy within the signal feature, and we show that the compressed signal features are competitive in distinguishing the signal compared to the original features. Second, we show by experiment that the neural network-based classification model almost always outperforms nonneural methods such as logistic regression, support vector machines, decision trees, and boosted trees.
ISSN
2375-933X
URI
https://hdl.handle.net/10371/186651
DOI
https://doi.org/10.1109/BIGCOMP.2019.8679115
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share