ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2510.21285
  4. Cited By
When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large Reasoning Models with Chain-of-Guardrails
v1v2 (latest)

When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large Reasoning Models with Chain-of-Guardrails

24 October 2025
Yingzhi Mao
Chunkang Zhang
Junxiang Wang
Xinyan Guan
Boxi Cao
Yaojie Lu
Hongyu Lin
Xianpei Han
Le Sun
    LRMELM
ArXiv (abs)PDFHTMLHuggingFace (4 upvotes)Github (2★)

Papers citing "When Models Outthink Their Safety: Mitigating Self-Jailbreak in Large Reasoning Models with Chain-of-Guardrails"

0 / 0 papers shown
Title

No papers found