ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2501.01830
  4. Cited By
Auto-RT: Automatic Jailbreak Strategy Exploration for Red-Teaming Large Language Models

Auto-RT: Automatic Jailbreak Strategy Exploration for Red-Teaming Large Language Models

3 January 2025
Yanjiang Liu
Shuhen Zhou
Yaojie Lu
Huijia Zhu
Weiqiang Wang
Hongyu Lin
Xianpei Han
Jia Zheng
Le Sun
ArXivPDFHTML

Papers citing "Auto-RT: Automatic Jailbreak Strategy Exploration for Red-Teaming Large Language Models"

1 / 1 papers shown
Title
Lifelong Safety Alignment for Language Models
Lifelong Safety Alignment for Language Models
Haoyu Wang
Zeyu Qin
Yifei Zhao
C. Du
Min Lin
Xueqian Wang
Tianyu Pang
KELM
CLL
56
1
0
26 May 2025
1