ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2308.02122
  4. Cited By
ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned
  Samples in NLP

ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP

4 August 2023
Lu Yan
Zhuo Zhang
Guanhong Tao
Kaiyuan Zhang
Xuan Chen
Guangyu Shen
Xiangyu Zhang
    AAML
    SILM
ArXivPDFHTML

Papers citing "ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP"

4 / 4 papers shown
Title
When LLM Meets DRL: Advancing Jailbreaking Efficiency via DRL-guided Search
When LLM Meets DRL: Advancing Jailbreaking Efficiency via DRL-guided Search
Xuan Chen
Yuzhou Nie
Wenbo Guo
Xiangyu Zhang
112
10
0
28 Jan 2025
On the Challenges of Fuzzing Techniques via Large Language Models
On the Challenges of Fuzzing Techniques via Large Language Models
Linghan Huang
Peizhou Zhao
Huaming Chen
Lei Ma
16
14
0
01 Feb 2024
Mind the Style of Text! Adversarial and Backdoor Attacks Based on Text
  Style Transfer
Mind the Style of Text! Adversarial and Backdoor Attacks Based on Text Style Transfer
Fanchao Qi
Yangyi Chen
Xurui Zhang
Mukai Li
Zhiyuan Liu
Maosong Sun
AAML
SILM
82
175
0
14 Oct 2021
Mitigating backdoor attacks in LSTM-based Text Classification Systems by
  Backdoor Keyword Identification
Mitigating backdoor attacks in LSTM-based Text Classification Systems by Backdoor Keyword Identification
Chuanshuai Chen
Jiazhu Dai
SILM
55
126
0
11 Jul 2020
1