ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.23559
  4. Cited By
SafeScientist: Toward Risk-Aware Scientific Discoveries by LLM Agents

SafeScientist: Toward Risk-Aware Scientific Discoveries by LLM Agents

29 May 2025
Kunlun Zhu
Jiaxun Zhang
Ziheng Qi
Nuoxing Shang
Zijia Liu
Peixuan Han
Yue Su
Haofei Yu
Jiaxuan You
ArXiv (abs)PDFHTML

Papers citing "SafeScientist: Toward Risk-Aware Scientific Discoveries by LLM Agents"

6 / 6 papers shown
Title
Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?
Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?
Yoshua Bengio
Michael K. Cohen
Damiano Fornasiere
J. Ghosn
Pietro Greiner
...
Jesse Richardson
Oliver E. Richardson
Marc-Antoine Rondeau
P. St-Charles
David Williams-King
82
18
0
21 Feb 2025
Red-Teaming LLM Multi-Agent Systems via Communication Attacks
Red-Teaming LLM Multi-Agent Systems via Communication Attacks
Pengfei He
Yupin Lin
Shen Dong
Han Xu
Yue Xing
Hui Liu
LLMAGAAML
94
9
0
20 Feb 2025
Agent Laboratory: Using LLM Agents as Research Assistants
Agent Laboratory: Using LLM Agents as Research Assistants
Samuel Schmidgall
Yusheng Su
Zihan Wang
Xingwu Sun
Jialian Wu
Xiaodong Yu
Jiang Liu
Michael Moor
Zicheng Liu
Emad Barsoum
LLMAG
94
60
2
08 Jan 2025
Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based Agents
H. Zhang
Jingyuan Huang
Kai Mei
Yifei Yao
Zhenting Wang
Chenlu Zhan
Hongwei Wang
Yongfeng Zhang
AAMLLLMAGELM
189
40
0
03 Oct 2024
SORRY-Bench: Systematically Evaluating Large Language Model Safety Refusal
SORRY-Bench: Systematically Evaluating Large Language Model Safety Refusal
Tinghao Xie
Xiangyu Qi
Yi Zeng
Yangsibo Huang
Udari Madhushani Sehwag
...
Bo Li
Kai Li
Danqi Chen
Peter Henderson
Prateek Mittal
ALMELM
182
79
0
20 Jun 2024
Certifying LLM Safety against Adversarial Prompting
Certifying LLM Safety against Adversarial Prompting
Aounon Kumar
Chirag Agarwal
Suraj Srinivas
Aaron Jiaxun Li
Soheil Feizi
Himabindu Lakkaraju
AAML
147
196
0
06 Sep 2023
1