期刊文献+
共找到288篇文章
< 1 2 15 >
每页显示 20 50 100
A survey of backdoor attacks and defenses:From deep neural networks to large language models
1
作者 Ling-Xin Jin Wei Jiang +5 位作者 Xiang-Yu Wen Mei-Yu Lin Jin-Yu Zhan Xing-Zhi Zhou Maregu Assefa Habtie Naoufel Werghi 《Journal of Electronic Science and Technology》 2025年第3期13-35,共23页
Deep neural networks(DNNs)have found extensive applications in safety-critical artificial intelligence systems,such as autonomous driving and facial recognition systems.However,recent research has revealed their susce... Deep neural networks(DNNs)have found extensive applications in safety-critical artificial intelligence systems,such as autonomous driving and facial recognition systems.However,recent research has revealed their susceptibility to backdoors maliciously injected by adversaries.This vulnerability arises due to the intricate architecture and opacity of DNNs,resulting in numerous redundant neurons embedded within the models.Adversaries exploit these vulnerabilities to conceal malicious backdoor information within DNNs,thereby causing erroneous outputs and posing substantial threats to the efficacy of DNN-based applications.This article presents a comprehensive survey of backdoor attacks against DNNs and the countermeasure methods employed to mitigate them.Initially,we trace the evolution of the concept from traditional backdoor attacks to backdoor attacks against DNNs,highlighting the feasibility and practicality of generating backdoor attacks against DNNs.Subsequently,we provide an overview of notable works encompassing various attack and defense strategies,facilitating a comparative analysis of their approaches.Through these discussions,we offer constructive insights aimed at refining these techniques.Finally,we extend our research perspective to the domain of large language models(LLMs)and synthesize the characteristics and developmental trends of backdoor attacks and defense methods targeting LLMs.Through a systematic review of existing studies on backdoor vulnerabilities in LLMs,we identify critical open challenges in this field and propose actionable directions for future research. 展开更多
关键词 backdoor Attacks backdoor defenses Deep neural networks Large language model
在线阅读 下载PDF
Proactive Disentangled Modeling of Trigger-Object Pairings for Backdoor Defense
2
作者 Kyle Stein Andrew AMahyari +1 位作者 Guillermo Francia III Eman El-Sheikh 《Computers, Materials & Continua》 2025年第10期1001-1018,共18页
Deep neural networks(DNNs)and generative AI(GenAI)are increasingly vulnerable to backdoor attacks,where adversaries embed triggers into inputs to cause models to misclassify or misinterpret target labels.Beyond tradit... Deep neural networks(DNNs)and generative AI(GenAI)are increasingly vulnerable to backdoor attacks,where adversaries embed triggers into inputs to cause models to misclassify or misinterpret target labels.Beyond traditional single-trigger scenarios,attackers may inject multiple triggers across various object classes,forming unseen backdoor-object configurations that evade standard detection pipelines.In this paper,we introduce DBOM(Disentangled Backdoor-Object Modeling),a proactive framework that leverages structured disentanglement to identify and neutralize both seen and unseen backdoor threats at the dataset level.Specifically,DBOM factorizes input image representations by modeling triggers and objects as independent primitives in the embedding space through the use of Vision-Language Models(VLMs).By leveraging the frozen,pre-trained encoders of VLMs,our approach decomposes the latent representations into distinct components through a learnable visual prompt repository and prompt prefix tuning,ensuring that the relationships between triggers and objects are explicitly captured.To separate trigger and object representations in the visual prompt repository,we introduce the trigger–object separation and diversity losses that aids in disentangling trigger and object visual features.Next,by aligning image features with feature decomposition and fusion,as well as learned contextual prompt tokens in a shared multimodal space,DBOM enables zero-shot generalization to novel trigger-object pairings that were unseen during training,thereby offering deeper insights into adversarial attack patterns.Experimental results on CIFAR-10 and GTSRB demonstrate that DBOM robustly detects poisoned images prior to downstream training,significantly enhancing the security of DNN training pipelines. 展开更多
关键词 backdoor attacks generative AI DISENTANGLEMENT
在线阅读 下载PDF
Defending against Backdoor Attacks in Federated Learning by Using Differential Privacy and OOD Data Attributes
3
作者 Qingyu Tan Yan Li Byeong-Seok Shin 《Computer Modeling in Engineering & Sciences》 2025年第5期2417-2428,共12页
Federated Learning(FL),a practical solution that leverages distributed data across devices without the need for centralized data storage,which enables multiple participants to jointly train models while preserving dat... Federated Learning(FL),a practical solution that leverages distributed data across devices without the need for centralized data storage,which enables multiple participants to jointly train models while preserving data privacy and avoiding direct data sharing.Despite its privacy-preserving advantages,FL remains vulnerable to backdoor attacks,where malicious participants introduce backdoors into local models that are then propagated to the global model through the aggregation process.While existing differential privacy defenses have demonstrated effectiveness against backdoor attacks in FL,they often incur a significant degradation in the performance of the aggregated models on benign tasks.To address this limitation,we propose a novel backdoor defense mechanism based on differential privacy.Our approach first utilizes the inherent out-of-distribution characteristics of backdoor samples to identify and exclude malicious model updates that significantly deviate from benign models.By filtering out models that are clearly backdoor-infected before applying differential privacy,our method reduces the required noise level for differential privacy,thereby enhancing model robustness while preserving performance.Experimental evaluations on the CIFAR10 and FEMNIST datasets demonstrate that our method effectively limits the backdoor accuracy to below 15%across various backdoor scenarios while maintaining high main task accuracy. 展开更多
关键词 Federated learning backdoor attacks differential privacy out-of-distribution data
在线阅读 下载PDF
How Robust Are Language Models against Backdoors in Federated Learning?
4
作者 Seunghan Kim Changhoon Lim +1 位作者 Gwonsang Ryu Hyunil Kim 《Computer Modeling in Engineering & Sciences》 2025年第11期2617-2630,共14页
Federated Learning enables privacy-preserving training of Transformer-based language models,but remains vulnerable to backdoor attacks that compromise model reliability.This paper presents a comparative analysis of de... Federated Learning enables privacy-preserving training of Transformer-based language models,but remains vulnerable to backdoor attacks that compromise model reliability.This paper presents a comparative analysis of defense strategies against both classical and advanced backdoor attacks,evaluated across autoencoding and autoregressive models.Unlike prior studies,this work provides the first systematic comparison of perturbation-based,screening-based,and hybrid defenses in Transformer-based FL environments.Our results show that screening-based defenses consistently outperform perturbation-based ones,effectively neutralizing most attacks across architectures.However,this robustness comes with significant computational overhead,revealing a clear trade-off between security and efficiency.By explicitly identifying this trade-off,our study advances the understanding of defense strategies in federated learning and highlights the need for lightweight yet effective screening methods for trustworthy deployment in diverse application domains. 展开更多
关键词 backdoor attack federated learning transformer-based language model system robustness
在线阅读 下载PDF
Poison-Only and Targeted Backdoor Attack Against Visual Object Tracking
5
作者 GU Wei SHAO Shuo +2 位作者 ZHOU Lingtao QIN Zhan REN Kui 《ZTE Communications》 2025年第3期3-14,共12页
Visual object tracking(VOT),aiming to track a target object in a continuous video,is a fundamental and critical task in computer vision.However,the reliance on third-party resources(e.g.,dataset)for training poses con... Visual object tracking(VOT),aiming to track a target object in a continuous video,is a fundamental and critical task in computer vision.However,the reliance on third-party resources(e.g.,dataset)for training poses concealed threats to the security of VOT models.In this paper,we reveal that VOT models are vulnerable to a poison-only and targeted backdoor attack,where the adversary can achieve arbitrary tracking predictions by manipulating only part of the training data.Specifically,we first define and formulate three different variants of the targeted attacks:size-manipulation,trajectory-manipulation,and hybrid attacks.To implement these,we introduce Random Video Poisoning(RVP),a novel poison-only strategy that exploits temporal correlations within video data by poisoning entire video sequences.Extensive experiments demonstrate that RVP effectively injects controllable backdoors,enabling precise manipulation of tracking behavior upon trigger activation,while maintaining high performance on benign data,thus ensuring stealth.Our findings not only expose significant vulnerabilities but also highlight that the underlying principles could be adapted for beneficial uses,such as dataset watermarking for copyright protection. 展开更多
关键词 visual object tracking backdoor attack computer vision data security AI safety
在线阅读 下载PDF
基于剪枝与后门遗忘的深度神经网络后门移除方法
6
作者 李学相 高亚飞 +2 位作者 夏辉丽 王超 刘明林 《郑州大学学报(工学版)》 北大核心 2026年第2期27-34,共8页
后门攻击对深度神经网络的安全性构成了严重威胁。现有的大多数后门防御方法依赖部分原始训练数据来移除模型中的后门,但在数据访问受限这一现实场景中,这些方法在移除模型后门时的效果不佳,并且对模型的原始精度产生较大影响。针对上... 后门攻击对深度神经网络的安全性构成了严重威胁。现有的大多数后门防御方法依赖部分原始训练数据来移除模型中的后门,但在数据访问受限这一现实场景中,这些方法在移除模型后门时的效果不佳,并且对模型的原始精度产生较大影响。针对上述问题,提出了一种基于剪枝和后门遗忘的无数据后门移除方法(DBR-PU)。首先,用所提方法分析模型神经元在合成数据集上的预激活分布差异,以此来定位可疑神经元;其次,通过对这些可疑神经元进行剪枝操作来降低后门对模型的影响;最后,使用对抗性后门遗忘策略来进一步消除模型对少量残留后门信息的内部响应。在CIFAR10和GTSRB数据集上对6种主流后门攻击方法进行实验,结果表明:在数据访问受限的条件下,所提方法在准确率上可以与最优的基准防御方法保持较小差距,并且在降低攻击成功率方面表现最好。 展开更多
关键词 深度神经网络 后门攻击 后门防御 预激活分布 对抗性后门遗忘
在线阅读 下载PDF
基于奇异值空间对抗优化的动态后门攻击
7
作者 彭子铭 丁建伟 +1 位作者 姚佳旺 田华伟 《计算机科学与探索》 北大核心 2026年第2期510-521,共12页
深度学习模型凭借其卓越的性能已在众多领域得到广泛应用,但研究表明其对后门攻击也具有显著的脆弱性。后门攻击可通过隐蔽的触发机制破坏模型的可靠性,当预设的触发器激活隐藏后门时,模型将执行恶意行为。目前后门攻击主要依赖于空间... 深度学习模型凭借其卓越的性能已在众多领域得到广泛应用,但研究表明其对后门攻击也具有显著的脆弱性。后门攻击可通过隐蔽的触发机制破坏模型的可靠性,当预设的触发器激活隐藏后门时,模型将执行恶意行为。目前后门攻击主要依赖于空间域或频域的扰动触发模式,且多采用样本无关的静态触发器设置,使得防御系统能够相对容易地检测并消除威胁。为了解决现有攻击隐蔽性不足和鲁棒性较弱的问题,提出一种基于奇异值空间进行阶段性对抗优化的动态后门攻击方法。通过生成器生成具有样本特异性的触发器,利用奇异值分解(SVD)提取干净图像和触发器的主/次特征,将触发信息嵌入干净图像次特征中,保留主特征以维持后门隐蔽性。提出阶段性训练框架:第一阶段联合优化触发生成器与分类器,确保最大化后门攻击的有效性;第二阶段则用最优触发生成器继续训练后门模型。为了验证方法的隐蔽性与有效性,在多个经典数据集上测试了攻击方法。实验结果表明,该方法在四个数据集上都实现了比现有攻击方法更高的攻击成功率,且在良性样本上几乎没有导致准确率下降,并绕过了四种先进的后门防御方法。同时,实验还验证了深度模型对奇异值扰动的敏感性可被恶意利用,而现有的防御机制难以识别此类攻击,为AI模型揭示了新的安全隐患。 展开更多
关键词 后门攻击 阶段性对抗优化 奇异值分解 样本特异性 模型安全
在线阅读 下载PDF
基于特征调控与颜色分离的净标签多后门攻击方法
8
作者 唐迎春 黄荣 +1 位作者 周树波 蒋学芹 《计算机应用》 北大核心 2026年第1期124-134,共11页
针对传统的后门攻击缺乏隐蔽性与灵活性的问题,提出一种基于特征调控与颜色分离的净标签多后门攻击方法,以信息隐藏框架为基础,训练中毒网络嵌入触发器。首先,以图像边缘作为触发器,设计特征调控策略,结合对抗扰动与代理模型辅助训练中... 针对传统的后门攻击缺乏隐蔽性与灵活性的问题,提出一种基于特征调控与颜色分离的净标签多后门攻击方法,以信息隐藏框架为基础,训练中毒网络嵌入触发器。首先,以图像边缘作为触发器,设计特征调控策略,结合对抗扰动与代理模型辅助训练中毒网络,增强触发器特征的显著性;其次,提出颜色分离策略对触发器进行着色,赋予触发器可区分的RGB空间颜色并设置与颜色相对应的one-hot目标置信度引导训练,从而保证触发器特征的可区分性。为了验证所提方法的有效性,分别在3个数据集(CIFAR-10、ImageNet-10和GTSRB)上以及5种模型上进行实验。结果表明,在单后门场景下,所提方法的攻击成功率(ASR)在3个数据集上均超过98%,分别超过次优方法7.94、1.70和8.61个百分点;在多后门场景下,所提方法在ImageNet-10数据集上的ASR达到90%以上,平均ASR超过次优方法36.63个百分点。而消融实验的结果也验证了特征调控与颜色分离策略的合理性及对抗扰动与代理模型的贡献,多后门实验的结果展示了所提攻击方法的灵活性。 展开更多
关键词 后门攻击 净标签 特征调控 颜色分离 代理模型
在线阅读 下载PDF
微波参数对金属腔体强耦合覆盖率的影响研究
9
作者 曹垒 冯溪溪 +3 位作者 陈自东 戈弋 赵刚 赵景涛 《强激光与粒子束》 北大核心 2026年第1期73-81,共9页
针对精确评估目标高功率微波后门耦合效应困难的问题,以典型金属腔体为对象,提出“强耦合区域覆盖率”指标,采用时域有限差分法(FDTD)建立仿真模型,分析了高功率微波波形参数对后门耦合效应的影响。研究发现:腔体固有谐振频率下强耦合... 针对精确评估目标高功率微波后门耦合效应困难的问题,以典型金属腔体为对象,提出“强耦合区域覆盖率”指标,采用时域有限差分法(FDTD)建立仿真模型,分析了高功率微波波形参数对后门耦合效应的影响。研究发现:腔体固有谐振频率下强耦合覆盖率显著高于非谐振频率;脉冲宽度增至特定值时,谐振频率下覆盖率趋于稳定;极化角从水平向垂直变化时,覆盖率有显著变化;叠加不同谐振频率可填补非强耦合区域空白,进一步提高覆盖率;脉冲前沿对覆盖率影响较小。该研究可为高功率微波后门耦合效应机理及参数优化提供关键技术支撑。 展开更多
关键词 强耦合区域覆盖率 金属腔体 时域有限差分法 高功率微波参数 后门耦合效应
在线阅读 下载PDF
分离触发器和多重对比的数据浓缩后门攻击
10
作者 蒋桂政 黄荣 +1 位作者 刘浩 蒋学芹 《中国图象图形学报》 北大核心 2026年第1期177-196,共20页
目的现有数据浓缩后门攻击方法将含有触发器的中毒样本和干净样本浓缩为小的数据集,中毒数据中真实数据的强信号掩盖触发器的弱信号,并且未考虑将非目标类浓缩数据与中毒数据特征分离,非目标类浓缩数据残留触发器特征。因此,提出分离触... 目的现有数据浓缩后门攻击方法将含有触发器的中毒样本和干净样本浓缩为小的数据集,中毒数据中真实数据的强信号掩盖触发器的弱信号,并且未考虑将非目标类浓缩数据与中毒数据特征分离,非目标类浓缩数据残留触发器特征。因此,提出分离触发器和多重对比的数据浓缩后门攻击。方法首先将触发器与真实数据进行分离。分离的触发器作为样本与真实数据并行嵌入浓缩数据,减少真实数据对触发器的干扰。然后,对分离的触发器进行优化,将触发器接近目标类真实数据的特征,提高触发器的嵌入效果,同时对触发器进行了分区放大预处理来增加触发器像素的数量,使其在优化过程获取大量的梯度用于指导学习。在数据浓缩阶段,通过多重对比将目标类浓缩数据与触发器特征投影在同一空间,将非目标类浓缩数据与触发器特征分离,进一步提高后门攻击的成功率。结果为了验证所提出方法的有效性,将所提出方法在FashionMNIST(Fashion Modified National Institute of Standards and Technology database)、CIFAR10(Canadian Institute for Advances Research’s ten categories dataset)、STL10(Stanford letter-10)、SVHN(street view house numbers)与其他4种方法进行对比实验。所提出的方法在5个数据集和6个不同的模型上均达到100%的攻击成功率,同时未降低干净样本在模型上的准确率。结论所提出的方法通过解决现有方法存在的问题,实现了性能的显著提高。本文方法具体代码见:https://github.com/tfuy/STMC。 展开更多
关键词 后门攻击 数据浓缩 分离 梯度匹配 分区放大预处理 最大化
原文传递
基于自监督学习与数据集分割的后门防御方法
11
作者 何子晟 凌捷 《计算机应用研究》 北大核心 2026年第1期256-262,共7页
针对深度神经网络(DNNs)在图像分类任务中易受后门攻击、现有防御方法难以兼顾模型准确率与鲁棒性的问题,提出一种名为SAS(self-supervised adaptive splitting)的基于自监督预训练与动态数据集分割的半监督后门防御方法。该方法首先引... 针对深度神经网络(DNNs)在图像分类任务中易受后门攻击、现有防御方法难以兼顾模型准确率与鲁棒性的问题,提出一种名为SAS(self-supervised adaptive splitting)的基于自监督预训练与动态数据集分割的半监督后门防御方法。该方法首先引入一致性正则化的对比学习框架进行自监督训练,解耦图像特征与后门模式;随后的微调阶段基于动态数据筛选与半监督学习策略,在训练中筛选并分别利用高可信度和低可信度数据,抑制后门植入。在CIFAR-10和GTSRB两种数据集上,针对BadNets、Blend、WaNet和Refool四种攻击的实验表明,该方法相较ASD方法,在两种数据集的干净数据上的分类准确率分别平均提升了1.65和0.65个百分点;污染数据的后门攻击成功率均降低到1.4%以下。研究证实,该方法通过解耦特征与动态数据集分割的协同作用,能有效提升模型的后门防御能力,同时保持在干净数据上的高分类性能,为构建安全可靠的深度学习模型提供了有效的途径。 展开更多
关键词 深度学习 后门防御 半监督学习 图像分类 自监督学习
在线阅读 下载PDF
基于变分自编码器触发器逆向的后门检测方法
12
作者 古津榜 洪征 +3 位作者 秦素娟 赵敏 陈乾 江川 《计算机工程与应用》 北大核心 2026年第2期245-256,共12页
针对现有触发器逆向方法难以检测出复杂后门攻击的问题,提出一种基于变分自编码器(VAE)触发器逆向的后门检测方法。使用变分自编码器作为空间变换函数,变分自编码器可以泛化出不同形式的触发器,干净样本经过变分自编码器处理后,被嵌入... 针对现有触发器逆向方法难以检测出复杂后门攻击的问题,提出一种基于变分自编码器(VAE)触发器逆向的后门检测方法。使用变分自编码器作为空间变换函数,变分自编码器可以泛化出不同形式的触发器,干净样本经过变分自编码器处理后,被嵌入触发器得到逆向样本。后门攻击具有隐蔽的特点,能够实施后门攻击的逆向样本往往需要与干净样本保持结构相似性,基于此特点选择满足隐蔽性要求的逆向样本。后门往往需要学习触发器特征,与指定的类别建立强联系。采用余弦相似性衡量逆向样本与干净样本的特征值之间的关系,来确定可能存在后门的逆向样本。在此基础上,应用L1范式约束优化触发器,使触发器在保持有效性的同时最小化L1范数以减小触发器对干净样本造成的影响,进而依据逆向样本与特定类别的强联系进行后门检测。实验结果表明,该方法可以有效识别不同特征空间中的复杂后门攻击。 展开更多
关键词 神经网络安全 后门攻击 触发器逆向 触发器 后门攻击检测
在线阅读 下载PDF
联合视觉-文本特征的复合型触发器后门攻击
13
作者 黄荣 唐迎春 +1 位作者 周树波 蒋学芹 《计算机科学》 北大核心 2026年第1期382-394,共13页
后门攻击指攻击者通过毒化数据集,隐蔽地诱导受害模型关联中毒数据和目标标签,对人工智能技术的可信和安全产生威胁。现有后门攻击方法普遍存在着有效性和隐蔽性之间顾此失彼的矛盾,有效性强的触发器隐蔽性差,反之,隐蔽性好的触发器有... 后门攻击指攻击者通过毒化数据集,隐蔽地诱导受害模型关联中毒数据和目标标签,对人工智能技术的可信和安全产生威胁。现有后门攻击方法普遍存在着有效性和隐蔽性之间顾此失彼的矛盾,有效性强的触发器隐蔽性差,反之,隐蔽性好的触发器有效性弱。针对该问题,提出一种联合视觉-文本特征的复合型触发器净标签后门攻击。复合型触发器由通用型和个性化两部分可学习的触发器叠加而成。复合型触发器的设计和优化均以块内像素值的同余为约束,旨在诱导受害模型捕捉同余规律,建立起触发器和目标标签的关联,形成后门。通用型触发器使得中毒图像的块内像素值对位权2同余,其信号形态对于所有的中毒图像单一固定;个性化触发器使得中毒图像的边缘像素值对LoSB(Lower Significant Bit)的位权同余,其信号特定于图像的边缘位置。两部分触发器相叠加,有利于兼顾有效性和隐蔽性。在此基础上,引入CLIP(Contrastive Language-Image Pre-training)模型,联合视觉和文本特征构建驱动复合型触发器训练的监督信号。预训练的CLIP模型具有较强的泛化能力,能够引导复合型触发器吸收异类的文本特征,起到弱化图像内容特征的作用,进一步增强触发器的有效性。在CIFAR-10,ImageNet,GTSRB这3个数据集上开展了实验。结果表明,所提方法能够抵御后门防御技术的侦测,在攻击成功率指标上平均超越次优方法2.48个百分点;在峰值信噪比、结构相似性度量、梯度幅度相似性偏差和学习感知图像块相似度4项指标上分别平均超越次优方法10.61%,0.31%,68.44%和46.38%。消融实验的结果验证了联合视觉和本文特征引导复合型触发器训练的优势,还验证了通用型和个性化两部分触发器对后门攻击的有效性和隐蔽性。 展开更多
关键词 后门攻击 复合型触发器 同余规律 CLIP模型
在线阅读 下载PDF
Software Backdoor Analysis Based on Sensitive Flow Tracking and Concolic Execution 被引量:3
14
作者 XU Xin WANG Jiajie +2 位作者 CHENG Shaoyin ZHANG Tao JIANG Fan 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2016年第5期421-427,共7页
In order to effectively detect and analyze the backdoors this paper introduces a method named Backdoor Analysis based on Sensitive flow tracking and Concolic Execution(BASEC).BASEC uses sensitive flow tracking to ef... In order to effectively detect and analyze the backdoors this paper introduces a method named Backdoor Analysis based on Sensitive flow tracking and Concolic Execution(BASEC).BASEC uses sensitive flow tracking to effectively discover backdoor behaviors, such as stealing secret information and injecting evil data into system, with less false negatives. With concolic execution on predetermined path, the backdoor trigger condition can be extracted and analyzed to achieve high accuracy. BASEC has been implemented and experimented on several software backdoor samples widespread on the Internet, and over 90% of them can be detected. Compared with behavior-based and system-call-based detection methods, BASEC relies less on the historical sample collections, and is more effective in detecting software backdoors, especially those injected into software by modifying and recompiling source codes. 展开更多
关键词 software backdoor detection data flow tracking concolic execution malware detection
原文传递
An Improved Optimized Model for Invisible Backdoor Attack Creation Using Steganography 被引量:2
15
作者 Daniyal M.Alghazzawi Osama Bassam J.Rabie +1 位作者 Surbhi Bhatia Syed Hamid Hasan 《Computers, Materials & Continua》 SCIE EI 2022年第7期1173-1193,共21页
The Deep Neural Networks(DNN)training process is widely affected by backdoor attacks.The backdoor attack is excellent at concealing its identity in the DNN by performing well on regular samples and displaying maliciou... The Deep Neural Networks(DNN)training process is widely affected by backdoor attacks.The backdoor attack is excellent at concealing its identity in the DNN by performing well on regular samples and displaying malicious behavior with data poisoning triggers.The state-of-art backdoor attacks mainly follow a certain assumption that the trigger is sample-agnostic and different poisoned samples use the same trigger.To overcome this problem,in this work we are creating a backdoor attack to check their strength to withstand complex defense strategies,and in order to achieve this objective,we are developing an improved Convolutional Neural Network(ICNN)model optimized using a Gradient-based Optimization(GBO)(ICNN-GBO)algorithm.In the ICNN-GBO model,we are injecting the triggers via a steganography and regularization technique.We are generating triggers using a single-pixel,irregular shape,and different sizes.The performance of the proposed methodology is evaluated using different performance metrics such as Attack success rate,stealthiness,pollution index,anomaly index,entropy index,and functionality.When the CNN-GBO model is trained with the poisoned dataset,it will map the malicious code to the target label.The proposed scheme’s effectiveness is verified by the experiments conducted on both the benchmark datasets namely CIDAR-10 andMSCELEB 1M dataset.The results demonstrate that the proposed methodology offers significant defense against the conventional backdoor attack detection frameworks such as STRIP and Neutral cleanse. 展开更多
关键词 Convolutional neural network gradient-based optimization STEGANOGRAPHY backdoor attack and regularization attack
在线阅读 下载PDF
Adaptive Backdoor Attack against Deep Neural Networks 被引量:1
16
作者 Honglu He Zhiying Zhu Xinpeng Zhang 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第9期2617-2633,共17页
In recent years,the number of parameters of deep neural networks(DNNs)has been increasing rapidly.The training of DNNs is typically computation-intensive.As a result,many users leverage cloud computing and outsource t... In recent years,the number of parameters of deep neural networks(DNNs)has been increasing rapidly.The training of DNNs is typically computation-intensive.As a result,many users leverage cloud computing and outsource their training procedures.Outsourcing computation results in a potential risk called backdoor attack,in which a welltrained DNN would performabnormally on inputs with a certain trigger.Backdoor attacks can also be classified as attacks that exploit fake images.However,most backdoor attacks design a uniformtrigger for all images,which can be easilydetectedand removed.In this paper,we propose a novel adaptivebackdoor attack.We overcome this defect and design a generator to assign a unique trigger for each image depending on its texture.To achieve this goal,we use a texture complexitymetric to create a specialmask for eachimage,which forces the trigger tobe embedded into the rich texture regions.The trigger is distributed in texture regions,which makes it invisible to humans.Besides the stealthiness of triggers,we limit the range of modification of backdoor models to evade detection.Experiments show that our method is efficient in multiple datasets,and traditional detectors cannot reveal the existence of a backdoor. 展开更多
关键词 backdoor attack AI security DNN
在线阅读 下载PDF
Byzantine Robust Federated Learning Scheme Based on Backdoor Triggers 被引量:1
17
作者 Zheng Yang Ke Gu Yiming Zuo 《Computers, Materials & Continua》 SCIE EI 2024年第5期2813-2831,共19页
Federated learning is widely used to solve the problem of data decentralization and can provide privacy protectionfor data owners. However, since multiple participants are required in federated learning, this allows a... Federated learning is widely used to solve the problem of data decentralization and can provide privacy protectionfor data owners. However, since multiple participants are required in federated learning, this allows attackers tocompromise. Byzantine attacks pose great threats to federated learning. Byzantine attackers upload maliciouslycreated local models to the server to affect the prediction performance and training speed of the global model. Todefend against Byzantine attacks, we propose a Byzantine robust federated learning scheme based on backdoortriggers. In our scheme, backdoor triggers are embedded into benign data samples, and then malicious localmodels can be identified by the server according to its validation dataset. Furthermore, we calculate the adjustmentfactors of local models according to the parameters of their final layers, which are used to defend against datapoisoning-based Byzantine attacks. To further enhance the robustness of our scheme, each localmodel is weightedand aggregated according to the number of times it is identified as malicious. Relevant experimental data showthat our scheme is effective against Byzantine attacks in both independent identically distributed (IID) and nonindependentidentically distributed (non-IID) scenarios. 展开更多
关键词 Federated learning Byzantine attacks backdoor triggers
在线阅读 下载PDF
XMAM:X-raying models with a matrix to reveal backdoor attacks for federated learning 被引量:1
18
作者 Jianyi Zhang Fangjiao Zhang +3 位作者 Qichao Jin Zhiqiang Wang Xiaodong Lin Xiali Hei 《Digital Communications and Networks》 SCIE CSCD 2024年第4期1154-1167,共14页
Federated Learning(FL),a burgeoning technology,has received increasing attention due to its privacy protection capability.However,the base algorithm FedAvg is vulnerable when it suffers from so-called backdoor attacks... Federated Learning(FL),a burgeoning technology,has received increasing attention due to its privacy protection capability.However,the base algorithm FedAvg is vulnerable when it suffers from so-called backdoor attacks.Former researchers proposed several robust aggregation methods.Unfortunately,due to the hidden characteristic of backdoor attacks,many of these aggregation methods are unable to defend against backdoor attacks.What's more,the attackers recently have proposed some hiding methods that further improve backdoor attacks'stealthiness,making all the existing robust aggregation methods fail.To tackle the threat of backdoor attacks,we propose a new aggregation method,X-raying Models with A Matrix(XMAM),to reveal the malicious local model updates submitted by the backdoor attackers.Since we observe that the output of the Softmax layer exhibits distinguishable patterns between malicious and benign updates,unlike the existing aggregation algorithms,we focus on the Softmax layer's output in which the backdoor attackers are difficult to hide their malicious behavior.Specifically,like medical X-ray examinations,we investigate the collected local model updates by using a matrix as an input to get their Softmax layer's outputs.Then,we preclude updates whose outputs are abnormal by clustering.Without any training dataset in the server,the extensive evaluations show that our XMAM can effectively distinguish malicious local model updates from benign ones.For instance,when other methods fail to defend against the backdoor attacks at no more than 20%malicious clients,our method can tolerate 45%malicious clients in the black-box mode and about 30%in Projected Gradient Descent(PGD)mode.Besides,under adaptive attacks,the results demonstrate that XMAM can still complete the global model training task even when there are 40%malicious clients.Finally,we analyze our method's screening complexity and compare the real screening time with other methods.The results show that XMAM is about 10–10000 times faster than the existing methods. 展开更多
关键词 Federated learning backdoor attacks Aggregation methods
在线阅读 下载PDF
Backdoor Malware Detection in Industrial IoT Using Machine Learning
19
作者 Maryam Mahsal Khan Attaullah Buriro +1 位作者 Tahir Ahmad Subhan Ullah 《Computers, Materials & Continua》 SCIE EI 2024年第12期4691-4705,共15页
With the ever-increasing continuous adoption of Industrial Internet of Things(IoT)technologies,security concerns have grown exponentially,especially regarding securing critical infrastructures.This is primarily due to... With the ever-increasing continuous adoption of Industrial Internet of Things(IoT)technologies,security concerns have grown exponentially,especially regarding securing critical infrastructures.This is primarily due to the potential for backdoors to provide unauthorized access,disrupt operations,and compromise sensitive data.Backdoors pose a significant threat to the integrity and security of Industrial IoT setups by exploiting vulnerabilities and bypassing standard authentication processes.Hence its detection becomes of paramount importance.This paper not only investigates the capabilities of Machine Learning(ML)models in identifying backdoor malware but also evaluates the impact of balancing the dataset via resampling techniques,including Synthetic Minority Oversampling Technique(SMOTE),Synthetic Data Vault(SDV),and Conditional Tabular Generative Adversarial Network(CTGAN),and feature reduction such as Pearson correlation coefficient,on the performance of the ML models.Experimental evaluation on the CCCS-CIC-AndMal-2020 dataset demonstrates that the Random Forest(RF)classifier generated an optimal model with 99.98%accuracy when using a balanced dataset created by SMOTE.Additionally,the training and testing time was reduced by approximately 50%when switching from the full feature set to a reduced feature set,without significant performance loss. 展开更多
关键词 Industrial IoT backdoor malware machine learning CCCS-CIC-AndMal-2020 security detection critical infrastructure
在线阅读 下载PDF
A backdoor attack against quantum neural networks with limited information
20
作者 黄晨猗 张仕斌 《Chinese Physics B》 SCIE EI CAS CSCD 2023年第10期219-228,共10页
Backdoor attacks are emerging security threats to deep neural networks.In these attacks,adversaries manipulate the network by constructing training samples embedded with backdoor triggers.The backdoored model performs... Backdoor attacks are emerging security threats to deep neural networks.In these attacks,adversaries manipulate the network by constructing training samples embedded with backdoor triggers.The backdoored model performs as expected on clean test samples but consistently misclassifies samples containing the backdoor trigger as a specific target label.While quantum neural networks(QNNs)have shown promise in surpassing their classical counterparts in certain machine learning tasks,they are also susceptible to backdoor attacks.However,current attacks on QNNs are constrained by the adversary's understanding of the model structure and specific encoding methods.Given the diversity of encoding methods and model structures in QNNs,the effectiveness of such backdoor attacks remains uncertain.In this paper,we propose an algorithm that leverages dataset-based optimization to initiate backdoor attacks.A malicious adversary can embed backdoor triggers into a QNN model by poisoning only a small portion of the data.The victim QNN maintains high accuracy on clean test samples without the trigger but outputs the target label set by the adversary when predicting samples with the trigger.Furthermore,our proposed attack cannot be easily resisted by existing backdoor detection methods. 展开更多
关键词 backdoor attack quantum artificial intelligence security quantum neural network variational quantum circuit
原文传递
上一页 1 2 15 下一页 到第
使用帮助 返回顶部