43

DeepSight: An All-in-One LM Safety Toolkit

Bo Zhang
Jiaxuan Guo
Lijun Li
Dongrui Liu
Sujin Chen
Guanxu Chen
Zhijie Zheng
Qihao Lin
Lewen Yan
Chen Qian
Yijin Zhou
Yuyao Wu
Shaoxiong Guo
Tianyi Du
Jingyi Yang
Xuhao Hu
Ziqi Miao
Xiaoya Lu
Jing Shao
Xia Hu
Main:24 Pages
24 Figures
Bibliography:5 Pages
4 Tables
Abstract

As the development of Large Models (LMs) progresses rapidly, their safety is also a priority. In current Large Language Models (LLMs) and Multimodal Large Language Models (MLLMs) safety workflow, evaluation, diagnosis, and alignment are often handled by separate tools. Specifically, safety evaluation can only locate external behavioral risks but cannot figure out internal root causes. Meanwhile, safety diagnosis often drifts from concrete risk scenarios and remains at the explainable level. In this way, safety alignment lack dedicated explanations of changes in internal mechanisms, potentially degrading general capabilities. To systematically address these issues, we propose an open-source project, namely DeepSight, to practice a new safety evaluation-diagnosis integrated paradigm. DeepSight is low-cost, reproducible, efficient, and highly scalable large-scale model safety evaluation project consisting of a evaluation toolkit DeepSafe and a diagnosis toolkit DeepScan. By unifying task and data protocols, we build a connection between the two stages and transform safety evaluation from black-box to white-box insight. Besides, DeepSight is the first open source toolkit that support the frontier AI risk evaluation and joint safety evaluation and diagnosis.

View on arXiv
Comments on this paper