Fuzzy-based voiced-unvoiced segmentation for emotion recognition using spectral feature fusions

Despite abundant growth in automatic emotion recognition system (ERS) studies using various techniques in feature extractions and classifiers, scarce sources found to improve the system via pre-processing techniques. This paper proposed a smart pre-processing stage using fuzzy logic inference system...

Full description

Bibliographic Details
Published in:Indonesian Journal of Electrical Engineering and Computer Science
Main Author: Ali Y.M.; Rahim A.F.A.; Noorsal E.; Yassin Z.M.; Mokhtar N.F.; Ramlan M.H.
Format: Article
Language:English
Published: Institute of Advanced Engineering and Science 2020
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85083092765&doi=10.11591%2fijeecs.v19.i1.pp196-206&partnerID=40&md5=5849faa3152907cbeedf174173968813
Description
Summary:Despite abundant growth in automatic emotion recognition system (ERS) studies using various techniques in feature extractions and classifiers, scarce sources found to improve the system via pre-processing techniques. This paper proposed a smart pre-processing stage using fuzzy logic inference system (FIS) based on Mamdani engine and simple time-based features i.e. zero-crossing rate (ZCR) and short-time energy (STE) to initially identify a frame as voiced (V) or unvoiced (UV). Mel-frequency cepstral coefficients (MFCC) and linear prediction coefficients (LPC) were tested with K-nearest neighbours (KNN) classifiers to evaluate the proposed FIS V-UV segmentation. We also introduced two feature fusions of MFCC and LPC with formants to obtain better performance. Experimental results of the proposed system surpassed the conventional ERS which yielded a rise in accuracy rate from 3.7% to 9.0%. The fusion of LPC and formants named as SFF LPC-fmnt indicated a promising result between 1.3% and 5.1% higher accuracy rate than its baseline features in classifying between neutral, angry, happy and sad emotions. The best accuracy rates yielded for male and female speakers were 79.1% and 79.9% respectively using SFF MFCC-fmnt fusion technique. Copyright © 2020 Institute of Advanced Engineering and Science. All rights reserved.
ISSN:25024752
DOI:10.11591/ijeecs.v19.i1.pp196-206