PulseAugur
LIVE 06:16:04
research · [2 sources] ·
0
research

Researchers unveil new stealthy backdoor attacks on AI models using diffusion and style features

Researchers have developed new methods for backdoor attacks on advanced AI models, specifically targeting Vision-Language Models (VLMs) and Diffusion Models (DMs). One approach, CBV, uses diffusion models to create natural-looking poisoned examples for VLMs by subtly altering image generation processes and focusing modifications on semantically important regions. Another method, Gungnir, exploits stylistic features within images as stealthy triggers for diffusion models, making attacks harder to detect and bypass existing defenses. AI

Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →

IMPACT New attack vectors highlight vulnerabilities in VLMs and diffusion models, necessitating advancements in AI safety and defense mechanisms.

RANK_REASON Two research papers detailing novel backdoor attack methods on AI models.

Read on arXiv cs.AI →

COVERAGE [2]

  1. arXiv cs.AI TIER_1 · Ji Guo, Xiaolong Qin, Cencen Liu, Jielei Wang, Jierun Chen, Wenbo Jiang ·

    CBV: Clean-label Backdoor Attacks on Vision Language Models via Diffusion Models

    arXiv:2605.02202v1 Announce Type: new Abstract: Vision-Language Models (VLMs) have achieved remarkable success in tasks such as image captioning and visual question answering (VQA). However, as their applications become increasingly widespread, recent studies have revealed that V…

  2. arXiv cs.CV TIER_1 · Lei Zhang, Yu Pan, Bingrong Dai, Lin Wang ·

    Gungnir: Exploiting Stylistic Features in Images for Backdoor Attacks on Diffusion Models

    arXiv:2502.20650v5 Announce Type: replace Abstract: Diffusion Models (DMs) have achieved remarkable success in image generation, yet recent studies reveal their vulnerability to backdoor attacks, where adversaries manipulate outputs via covert triggers embedded in inputs. Existin…