ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.04234
35
0

Functional Homotopy: Smoothing Discrete Optimization via Continuous Parameters for LLM Jailbreak Attacks

5 October 2024
Zi Wang
Divyam Anshumaan
Ashish Hooda
Yudong Chen
Somesh Jha
    AAML
ArXivPDFHTML
Abstract

Optimization methods are widely employed in deep learning to identify and mitigate undesired model responses. While gradient-based techniques have proven effective for image models, their application to language models is hindered by the discrete nature of the input space. This study introduces a novel optimization approach, termed the \emph{functional homotopy} method, which leverages the functional duality between model training and input generation. By constructing a series of easy-to-hard optimization problems, we iteratively solve these problems using principles derived from established homotopy methods. We apply this approach to jailbreak attack synthesis for large language models (LLMs), achieving a 20%−30%20\%-30\%20%−30% improvement in success rate over existing methods in circumventing established safe open-source models such as Llama-2 and Llama-3.

View on arXiv
@article{wang2025_2410.04234,
  title={ Functional Homotopy: Smoothing Discrete Optimization via Continuous Parameters for LLM Jailbreak Attacks },
  author={ Zi Wang and Divyam Anshumaan and Ashish Hooda and Yudong Chen and Somesh Jha },
  journal={arXiv preprint arXiv:2410.04234},
  year={ 2025 }
}
Comments on this paper