期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
VOTI:Jailbreaking Vision-Language Models via Visual Obfuscation and Task Induction
1
作者 ZHU Yifan CHU Zhixuan REN Kui 《ZTE Communications》 2025年第3期15-26,共12页
In recent years,large vision-language models(VLMs)have achieved significant breakthroughs in cross-modal understanding and generation.However,the safety issues arising from their multimodal interactions become promine... In recent years,large vision-language models(VLMs)have achieved significant breakthroughs in cross-modal understanding and generation.However,the safety issues arising from their multimodal interactions become prominent.VLMs are vulnerable to jailbreak attacks,where attackers craft carefully designed prompts to bypass safety mechanisms,leading them to generate harmful content.To address this,we investigate the alignment between visual inputs and task execution,uncovering locality defects and attention biases in VLMs.Based on these findings,we propose VOTI,a novel jailbreak framework leveraging visual obfuscation and task induction.VOTI subtly embeds malicious keywords within neutral image layouts to evade detection,and breaks down harmful queries into a sequence of subtasks.This approach disperses malicious intent across modalities,exploiting VLMs’over-reliance on local visual cues and their fragility in multi-step reasoning to bypass global safety mechanisms.Implemented as an automated framework,VOTI integrates large language models as red-team assistants to generate and iteratively optimize jailbreak strategies.Extensive experiments across seven mainstream VLMs demonstrate VOTI’s effectiveness,achieving a 73.46%attack success rate on GPT-4o-mini.These results reveal critical vulnerabilities in VLMs,highlighting the urgent need for improving robust defenses and multimodal alignment. 展开更多
关键词 large vision-language models jailbreak attacks red teaming security of large models safety alignment
在线阅读 下载PDF
The ethical security of large language models:A systematic review
2
作者 Feng LIU Jiaqi JIANG +2 位作者 Yating LU Zhanyi HUANG Jiuming JIANG 《Frontiers of Engineering Management》 2025年第1期128-140,共13页
The widespread application of large language models(LLMs)has highlighted new security challenges and ethical concerns,attracting significant academic and societal attention.Analysis of the security vulnerabilities of ... The widespread application of large language models(LLMs)has highlighted new security challenges and ethical concerns,attracting significant academic and societal attention.Analysis of the security vulnerabilities of LLMs and their misuse in cybercrime reveals that their advanced text-generation capabilities pose serious threats to personal privacy,data security,and information integrity.In addition,the effectiveness of current LLM-based defense strategies has been reviewed and evaluated.This paper examines the social implications of LLMs and proposes future directions for enhancing their security applications and ethical governance,aiming to inform the development of the field. 展开更多
关键词 security of large language models ethical governance model defense adversarial training social impact
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部