4.6 Article

TWACapsNet: a capsule network with two-way attention mechanism for speech emotion recognition

相关参考文献

注意:仅列出部分参考文献,下载原文获取全部文献信息。
Article Computer Science, Cybernetics

Intelligent recognition system for viewpoint variations on gait and speech using CNN-CapsNet

G. Merlin Linda et al.

Summary: This research introduces an intelligent recognition system for viewpoint variations of gait and speech using a CNN-CapsNet model. It focuses on relative spatial hierarchies between gait features and challenges brought by speech variations. The proposed system aims to improve the performance in recognizing gait and speech variations.

INTERNATIONAL JOURNAL OF INTELLIGENT COMPUTING AND CYBERNETICS (2022)

Article Engineering, Mechanical

Imbalanced fault diagnosis of rolling bearing using improved MsR-GAN and feature enhancement-driven CapsNet

Jie Liu et al.

Summary: An improved fault diagnosis approach utilizing multi-scale residual GAN and feature enhancement-driven capsule network is proposed in this paper to address the issue of imbalanced fault data distribution. Experimental verification shows that this method performs well in handling imbalanced fault data, with better stability and diagnostic accuracy compared to current methods.

MECHANICAL SYSTEMS AND SIGNAL PROCESSING (2022)

Article Acoustics

GM-TCNet: Gated Multi-scale Temporal Convolutional Network using Emotion Causality for Speech Emotion Recognition*

Jia-Xin Ye et al.

Summary: In this paper, a Gated Multi-scale Temporal Convolutional Network (GM-TCNet) is proposed for Speech Emotion Recognition (SER), which utilizes a novel emotional causality representation learning component to capture the dynamics of emotion, leading to improved performance.

SPEECH COMMUNICATION (2022)

Article Computer Science, Information Systems

Exploiting the potentialities of features for speech emotion recognition

Dongdong Li et al.

Summary: Recent studies on speech signals have focused on emotional information and the importance of feature representation in speech emotion recognition (SER). Different combinations of features and models have a significant impact on SER performance, with the proposed ECFW method showing promising results in improving performance across different databases.

INFORMATION SCIENCES (2021)

Article Acoustics

Unsupervised feature selection and NMF de-noising for robust Speech Emotion Recognition

Surekha Reddy Bandela et al.

Summary: Speech feature fusion is commonly used for improving accuracy in Speech Emotion Recognition (SER), but it can increase complexity and processing time. This paper proposes unsupervised feature selection to overcome this issue, using various feature reduction techniques to enhance SER performance in both clean and noisy environments. The results show that the proposed SER system outperforms existing literature works, especially in noisy conditions, with the adoption of a dense Non-Negative Matrix Factorization (denseNMF) method for noise robustness.

APPLIED ACOUSTICS (2021)

Article Computer Science, Artificial Intelligence

MLT-DNet: Speech emotion recognition using 1D dilated CNN based on multi-learning trick approach

Mustaqeem et al.

Summary: The paper discusses an end-to-end real-time SER model based on DCNN, which extracts emotional features and learns long-term contextual dependencies through multi-learning strategy, achieving high recognition accuracy.

EXPERT SYSTEMS WITH APPLICATIONS (2021)

Article Acoustics

Speech Emotion Recognition Using Sequential Capsule Networks

Xixin Wu et al.

Summary: Researchers proposed a novel architecture of sequential capsule networks (CapNets) to preserve spatial information for better speech emotion recognition, and introduced a recurrent connection to capture contextual information. Experimental comparison showed the effectiveness of the proposed sequential CapNets.

IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING (2021)

Proceedings Paper Computer Science, Artificial Intelligence

The Application of Capsule Neural Network Based CNN for Speech Emotion Recognition

Xin-Cheng Wen et al.

Summary: The study introduces a model called CapCNN that combines CNN and CapsNet, which effectively handles speech emotion recognition tasks with balanced results on CASIA and EMODB datasets.

2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) (2021)

Article Computer Science, Information Systems

Computer-aided diagnosis (CAD) system based on multi-layer feature fusion network for skin lesion recognition in dermoscopy images

Ibtissam Bakkouri et al.

MULTIMEDIA TOOLS AND APPLICATIONS (2020)

Article Computer Science, Artificial Intelligence

CapsNet topology to classify tumours from brain images and comparative evaluation

Evgin Goceri

IET IMAGE PROCESSING (2020)

Article Acoustics

A novel feature selection method for speech emotion recognition

Turgut Ozseven

APPLIED ACOUSTICS (2019)

Article Engineering, Biomedical

Speech emotion recognition using deep 1D & 2D CNN LSTM networks

Jianfeng Zhao et al.

BIOMEDICAL SIGNAL PROCESSING AND CONTROL (2019)

Article Acoustics

Speech emotion recognition based on DNN-decision tree SVM model

Linhui Sun et al.

SPEECH COMMUNICATION (2019)

Article Engineering, Electrical & Electronic

Improving the performance of the speaker emotion recognition based on low dimension prosody features vector

Ashishkumar Prabhakar Gudmalwar et al.

INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY (2019)

Article Acoustics

Convolutional Neural Networks for Speech Recognition

Ossama Abdel-Hamid et al.

IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING (2014)

Article Computer Science, Artificial Intelligence

Spoken emotion recognition using hierarchical classifiers

Enrique M. Albornoz et al.

COMPUTER SPEECH AND LANGUAGE (2011)

Article Engineering, Electrical & Electronic

Combining Long Short-Term Memory and Dynamic Bayesian Networks for Incremental Emotion-Sensitive Artificial Listening

Martin Woellmer et al.

IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING (2010)

Article Computer Science, Interdisciplinary Applications

IEMOCAP: interactive emotional dyadic motion capture database

Carlos Busso et al.

LANGUAGE RESOURCES AND EVALUATION (2008)