摘要
In order to improve the efficiency of speech emotion recognition across corpora,a speech emotion transfer learning method based on the deep sparse auto-encoder is proposed.The algorithm first reconstructs a small amount of data in the target domain by training the deep sparse auto-encoder,so that the encoder can learn the low-dimensional structural representation of the target domain data.Then,the source domain data and the target domain data are coded by the trained deep sparse auto-encoder to obtain the reconstruction data of the low-dimensional structural representation close to the target domain.Finally,a part of the reconstructed tagged target domain data is mixed with the reconstructed source domain data to jointly train the classifier.This part of the target domain data is used to guide the source domain data.Experiments on the CASIA,SoutheastLab corpus show that the model recognition rate after a small amount of data transferred reached 89.2%and 72.4%on the DNN.Compared to the training results of the complete original corpus,it only decreased by 2%in the CASIA corpus,and only 3.4%in the SoutheastLab corpus.Experiments show that the algorithm can achieve the effect of labeling all data in the extreme case that the data set has only a small amount of data tagged.
为了提高跨语料库的语音情感识别效率,提出了一种基于深度稀疏自编码的语音情感迁移学习方法.算法首先通过训练深度稀疏自编码器来对目标域中的少量数据进行重建,使得编码器可以学习到目标域数据低维度的结构表征.然后,将源域数据和目标域数据通过训练好的深度稀疏自编码器,得到靠近目标域低维度的结构表征的重建数据.最后,利用部分重建的含标签的目标域数据与重建的源域数据混合后共同训练分类器,以便完成对源域数据的引导.在CASIA、SoutheastLab语料库上的实验表明,通过少量数据迁移后的模型识别率在DNN上达到了89.2%和72.4%.和完整原始语料库训练的结果相比,在CASIA上仅下降了2%,在SoutheastLab上仅下降了3.4%.实验说明,该算法能够在数据集只有少量数据有标签的极端情况下,达到逼近于所有数据都有标签的效果.
基金
The National Natural Science Foundation of China(No.61871213,61673108,61571106)
Six Talent Peaks Project in Jiangsu Province(No.2016-DZXX-023)