ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.19773
21
0

What Really Matters in Many-Shot Attacks? An Empirical Study of Long-Context Vulnerabilities in LLMs

26 May 2025
Sangyeop Kim
Yohan Lee
Yongwoo Song
Kimin Lee
    AAML
ArXiv (abs)PDFHTML
Main:8 Pages
21 Figures
Bibliography:5 Pages
12 Tables
Appendix:8 Pages
Abstract

We investigate long-context vulnerabilities in Large Language Models (LLMs) through Many-Shot Jailbreaking (MSJ). Our experiments utilize context length of up to 128K tokens. Through comprehensive analysis with various many-shot attack settings with different instruction styles, shot density, topic, and format, we reveal that context length is the primary factor determining attack effectiveness. Critically, we find that successful attacks do not require carefully crafted harmful content. Even repetitive shots or random dummy text can circumvent model safety measures, suggesting fundamental limitations in long-context processing capabilities of LLMs. The safety behavior of well-aligned models becomes increasingly inconsistent with longer contexts. These findings highlight significant safety gaps in context expansion capabilities of LLMs, emphasizing the need for new safety mechanisms.

View on arXiv
@article{kim2025_2505.19773,
  title={ What Really Matters in Many-Shot Attacks? An Empirical Study of Long-Context Vulnerabilities in LLMs },
  author={ Sangyeop Kim and Yohan Lee and Yongwoo Song and Kimin Lee },
  journal={arXiv preprint arXiv:2505.19773},
  year={ 2025 }
}
Comments on this paper