期刊文献+

基于分布式扰动的文本对抗训练方法

Textual Adversarial Training Method Based on Distributed Perturbation
下载PDF
导出
摘要 文本对抗防御旨在增强神经网络模型对不同对抗攻击的抵御能力,目前的文本对抗防御方法通常只能对某种特定对抗攻击有效,对于原理不同的对抗攻击效果甚微。为解决文本对抗防御方法的不足,提出一种文本对抗分布训练(TADT)方法,将TADT形式化为一个极小极大优化问题,其中内部最大化的目标是了解每个输入示例的对抗分布,外部最小化的目标是通过最小化预期损失来减小对抗示例的数量,并对基于梯度下降和同义词替换的攻击方法进行研究。在2个文本分类数据集上的实验结果表明,相比于DNE方法,在PWWS、GA、UAT等3种不同的对抗攻击下,TADT方法的准确率平均提升2%,相比于其他方法提升了10%以上,且在不影响干净样本准确率的前提下显著提升了模型的鲁棒性,并在各种对抗攻击下具有较高的准确率,展示了良好的泛化性能。 Text adversarial defense aims to enhance the resilience of neural network models against different adversarial attacks.The current text confrontation defense methods are usually only effective against certain specific confrontation attacks and have little effect on confrontation attacks with different principles.To address the deficiencies of existing textual adversarial defense methods and principles of adversarial attack methods,this paper proposes the Textual Adversarial Distribution Training(TADT)method and formalizes it as a minimax optimization problem.The goal of inner maximization is to learn the adversarial distribution of each input example.The goal of outer minimization is to reduce the number of adversarial examples by minimizing the expected loss.This paper mainly studies the attack method based on gradient descent and synonym replacement.The experimental results on two text classification datasets show that under three different adconfrontation attacks,Probability Weighted Word Saliency(PWWS),Genetic Attack(GA),and Unsupervised Adversarial Training(UAT),the accuracy of TADT is improved by an average of 2%compared with the latest Dirichlet Neighborhood Ensemble(DNE)method.Compared with other methods,the accuracy of TADT method is improved by more than 10%,and the accuracy of clean samples is not affected.On the premise of not affecting the accuracy of clean samples,TADT significantly improves the robustness of the model and has high accuracy under various adversarial attacks,showing good generalization performance.
作者 沈志东 岳恒宪 SHEN Zhidong;YUE Hengxian(School of Cyber Science and Engineering,Wuhan University,Wuhan 430000,China)
出处 《计算机工程》 CAS CSCD 北大核心 2023年第9期16-22,共7页 Computer Engineering
基金 国家重点研发计划(2018YFC1604000) 湖北省重点研发计划项目(2022BAA041)。
关键词 文本对抗分布 对抗训练 变分自动编码器 梯度下降 蒙特卡罗采样 textual adversarial distribution Adversarial Training(AT) variational autoencoder gradient descent Monte Carlo sampling
  • 相关文献

参考文献2

二级参考文献20

  • 1刘建峰,王琦,姚震球.基于神经网络的客货船造价估算[J].华东船舶工业学院学报,1996,10(2):23-28. 被引量:4
  • 2Rullen R V, Thorpe S J. Rate Coding Versus Temporal Order Coding:What the Retinal Ganglion Cells Tell the Visual Cortex [ J]. Neural Computation, 2001,13 ( 6 ) : 1255-1283.
  • 3Maass W. Networks of Spiking Neurons: The Third Generation of Neural Network Models [ J ]. Neural Networks, 1997,10 (9) : 1659-1671.
  • 4Maass W. Noisy Spiking Neurons with Temporal Coding Have More Computational Power than Sigmoidal Neurons[M]. Cambridge ,USA :MIT Press,1997:259-268.
  • 5徐彦.基于时间编码的Spiking神经网络有监督学习机制的研究[D].南京:河海大学,2013.
  • 6Bohte S M,Kok J N, La J A. Error-backpropagation in Temporally Encoded Networks of Spiking Neurons [ J ]. Neurocomputing, 2002,48 ( 1-4 ) : 17-37.
  • 7McKennoch S, Liu D, Bushnell L G. Fast Modifications of the Spikeprop Algorithm [ C ]//Proceedings of Inter- national Joint Conference on Neural Networks. Vancouver, Canada : IEEE Press, 2006 : 3970-3977.
  • 8Schrauwen B ,Campenhout J V. Extending Spikeprop [ C ]// Proceedings of International Joint Conference on NeuralNetworks. Budapest, Hungary .IEEE Press,2004:471-476.
  • 9Booij O, Nguyen H T. AGradient Descent Rule for Multiple Spiking Neurons Emitting Multiple Spikes [ J ]. Information Processing Letters, 2005,95 ( 6 ) : 552-558.
  • 10Ghosh-Dastidar S,Adeli H. ANew Supervised Learning Algorithm for Multiple Spiking Neural Networks with Application in Epilepsy and Seizure Detection [ J ]. Neural Networks ,2009,22 ( 2 ) : 1419-1431.

共引文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部