ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2412.00621
76
2

Exposing LLM Vulnerabilities: Adversarial Scam Detection and Performance

1 December 2024
Chen-Wei Chang
Shailik Sarkar
Shutonu Mitra
Qi Zhang
Hossein Salemi
Hemant Purohit
Fengxiu Zhang
Michin Hong
Jin-Hee Cho
Chang-Tien Lu
ArXivPDFHTML
Abstract

Can we trust Large Language Models (LLMs) to accurately predict scam? This paper investigates the vulnerabilities of LLMs when facing adversarial scam messages for the task of scam detection. We addressed this issue by creating a comprehensive dataset with fine-grained labels of scam messages, including both original and adversarial scam messages. The dataset extended traditional binary classes for the scam detection task into more nuanced scam types. Our analysis showed how adversarial examples took advantage of vulnerabilities of a LLM, leading to high misclassification rate. We evaluated the performance of LLMs on these adversarial scam messages and proposed strategies to improve their robustness.

View on arXiv
Comments on this paper