10

CHASM: Unveiling Covert Advertisements on Chinese Social Media

Jingyi Zheng
Tianyi Hu
Yule Liu
Zhen Sun
Zongmin Zhang
Zifan Peng
Wenhan Dong
Xinlei He
Main:10 Pages
9 Figures
Bibliography:4 Pages
10 Tables
Appendix:15 Pages
Abstract

Current benchmarks for evaluating large language models (LLMs) in social media moderation completely overlook a serious threat: covert advertisements, which disguise themselves as regular posts to deceive and mislead consumers into making purchases, leading to significant ethical and legal concerns. In this paper, we present the CHASM, a first-of-its-kind dataset designed to evaluate the capability of Multimodal Large Language Models (MLLMs) in detecting covert advertisements on social media. CHASM is a high-quality, anonymized, manually curated dataset consisting of 4,992 instances, based on real-world scenarios from the Chinese social media platform Rednote. The dataset was collected and annotated under strict privacy protection and quality control protocols. It includes many product experience sharing posts that closely resemble covert advertisements, making the dataset particularlythis http URLresults show that under both zero-shot and in-context learning settings, none of the current MLLMs are sufficiently reliable for detecting covertthis http URLfurther experiments revealed that fine-tuning open-source MLLMs on our dataset yielded noticeable performance gains. However, significant challenges persist, such as detecting subtle cues in comments and differences in visual and textualthis http URLprovide in-depth error analysis and outline future research directions. We hope our study can serve as a call for the research community and platform moderators to develop more precise defenses against this emerging threat.

View on arXiv
Comments on this paper