期刊文献+

基于唇重构与三维耦合CNN的多视角音唇一致性判别

Multi-View Lip Motion and Voice Consistency Judgment Based on Lip Reconstruction and Three-Dimensional Coupled CNN
下载PDF
导出
摘要 针对传统音唇一致性判别方法主要对正面唇动视频进行处理,未考虑视频采集角度变化对结果的影响,且容易忽略唇动过程中的时空特性等不足,文中以唇部角度变化对一致性判别的影响为研究重心,结合三维卷积神经网络在非线性表示和时空维度特征提取上的优势,提出了基于正面唇重构与三维耦合卷积神经网络的多视角音唇一致性判别方法。该方法先通过在生成器中引入自映射损失来提高正面重建效果,并采用基于自映射监督循环一致性生成对抗网络(SMS-CycleGAN)的唇重构方法对多视角唇图进行角度分类及正面重构;然后设计两个异构三维卷积神经网络,分别用来描述音频和视频信号,并提取包含长时时空关联信息的三维卷积特征;最后引入对比损失函数作为音视频信号匹配的相关度鉴别度量,将音视频网络输出耦合到同一表示空间,并进行一致性判别。实验结果表明,文中方法能重建出更高质量的正面唇图,一致性判别性能优于多种不同类型的比较方法。 The traditional consistency judgment methods of lip motion and voice mainly focus on processing the frontal lip motion video,without considering the impact of angle changes on the result during the video acquisition process.In addition,they are prone to ignoring the spatio-temporal characteristics of the lip movement process.Aiming at these problems,this paper focused on the influence of lip angle changes on consistency judgment,combined the advantages of three dimensional convolutional neural networks for non-linear representation and spatio-temporal dimensional feature extraction,and proposed a multi-view lip motion and voice consistency judgment method based on frontal lip reconstruction and three dimensional(3D)coupled convolutional neural network.Firstly,the self-mapping loss was introduced into the generator to improve the effect of frontal reconstruction,and then the lip reconstruction method based on self-mapping supervised cycle-consistent generative adversarial network(SMS-CycleGAN)was used for angle classification and frontal reconstruction of multi-view lip image.Secondly,two heterogeneous three dimensional convolution neural networks were designed to describe the audio and video signals respectively,and then the 3D convolution features containing long-term spatio-temporal correlation information were extracted.Finally,the contrastive loss function was introduced as the correlation discrimination measure of audio and video signal matching,and the output of the audio-video network was coupled into the same representation space for consistency judgment.The experimental results show that the method proposed in this paper can reconstruct frontal lip images of higher quality,and it is better than a variety of comparison methods on the performance of consistency judgment.
作者 朱铮宇 罗超 贺前华 彭炜锋 毛志炜 张顺四 ZHU Zhengyu;LUO Chao;HE Qianhua;PENG Weifeng;MAO Zhiwei;ZHANG Shunsi(Audio,Speech and Vision Processing Laboratory,South China University of Technology,Guangzhou 510640,Guangdong,China;School of Cyber Security,Guangdong Polytechnic Normal University,Guangzhou 510665,Guangdong,China;Guangzhou Quwan Network Technology Co.,Ltd.,Guangzhou 510665,Guangdong,China)
出处 《华南理工大学学报(自然科学版)》 EI CAS CSCD 北大核心 2023年第5期70-77,共8页 Journal of South China University of Technology(Natural Science Edition)
基金 国家自然科学基金资助项目(61672173) 国家重点研发计划项目(2018YFB1802200)。
关键词 一致性判别 生成对抗网络 卷积神经网络 正面重构 多模态 consistency judgment generative adversarial network convolutional neural network frontal recon-struction multi-modal
  • 相关文献

参考文献3

二级参考文献32

  • 1MI Faraj, J Bigun. S ynergy of lip-motion and acoustic features in biometric speech and speaker recognition[ J]. IEEE Transac- tions on Computer,2007,56(9): 1169- 1175. 被引量:1
  • 2S Kumagal, K Doman, et al. Detection of inconsistency between subject and speaker based on the co-occurrence of lip motion and voice towards speech scene extraction from news videos [ A]. IEEE International Symposium on Multimedia[ C]. Cali- fornia: IEEE,2011.311 - 318. 被引量:1
  • 3M Slaney,M Covell. Facesync:A linear operator for measuring synchronization of video facial images and audio track [ A ].Neural Information Processing Systems [ C ]. Denver: NIPSF, 2000. 814 - 820. 被引量:1
  • 4N Eveno, L Besacier. A speaker independent "liveness" test for audio-visual biomelrics [ A ]. Nineth European Conference on Speech Communication and Technology [ C ]. Lisbon: ISCA, 2005. 3081 - 3084. 被引量:1
  • 5G ChoUet, R Landais, et al. Some experiments in audio-visual speech processing [A ]. Non-Linear Speech Processing 2007 [ C]. Paris-ISCA, 2007.28 - 56. 被引量:1
  • 6A Sayo, Y Kajikawa, et al. Biometrics authentication method using lip motion in utterance[ A]. 8th International Conference on Information, Communications and Signal Processing [ C ]. Singapore: IEEF., 2011.1 - 5. 被引量:1
  • 7AA EL-Sallam, AS Mian. Correlation based speech-video syn- chronization[ J]. Pattern Recognition Letters, 2011,32 ( 6 ) : 780 - 786. 被引量:1
  • 8B Goswami, C Chan, et al. Speaker authentication using video- based lip information[ A]. IEEE, International Conference on A- coustics, Speech, and Signal Processing [ C ]. Prague: IEEE, 2011.1908 - 1910. 被引量:1
  • 9R Goecke, B MiUar. Statistical analysis of the relationship be- tween audio and video speech parameters for Australian Eng- lish[ A]. Auditory Visual Speech Processing Conference[ C]. France: ISCA,2003.133 - 138. 被引量:1
  • 10ME Sargin, Y Yemez, et al. Audiovisual synchronization and fusion using canonical correlation analysis[ J]. IEEE Transac- tions on Multimedia,2007,9(7) : 1396 - 1402. 被引量:1

共引文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部