Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2409.03363
Cited By
v1
v2 (latest)
Con-ReCall: Detecting Pre-training Data in LLMs via Contrastive Decoding
5 September 2024
Cheng Wang
Yiwei Wang
Bryan Hooi
Yujun Cai
Nanyun Peng
Kai-Wei Chang
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Con-ReCall: Detecting Pre-training Data in LLMs via Contrastive Decoding"
28 / 28 papers shown
Title
SOFT: Selective Data Obfuscation for Protecting LLM Fine-tuning against Membership Inference Attacks
Kaiyuan Zhang
Siyuan Cheng
Hanxi Guo
Yuetian Chen
Zian Su
...
Yuntao Du
Charles Fleming
Ashish Kundu
Xiangyu Zhang
Ninghui Li
AAML
144
0
0
12 Jun 2025
Strong Membership Inference Attacks on Massive Datasets and (Moderately) Large Language Models
Jamie Hayes
Ilia Shumailov
Christopher A. Choquette-Choo
Matthew Jagielski
G. Kaissis
...
Matthieu Meeus
Yves-Alexandre de Montjoye
Franziska Boenisch
Adam Dziedzic
A. Feder Cooper
58
1
0
24 May 2025
On Membership Inference Attacks in Knowledge Distillation
Ziyao Cui
Minxing Zhang
Jian Pei
73
0
0
17 May 2025
Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models
Haritz Puerto
Martin Gubri
Sangdoo Yun
Seong Joon Oh
MIALM
771
3
2
31 Oct 2024
ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Roy Xie
Junlin Wang
Ruomin Huang
Minxing Zhang
Rong Ge
Jian Pei
Neil Zhenqiang Gong
Bhuwan Dhingra
MIALM
128
17
0
23 Jun 2024
Enhancing Contextual Understanding in Large Language Models through Contrastive Decoding
Zheng Zhao
Emilio Monti
Jens Lehmann
H. Assem
97
33
0
04 May 2024
Min-K%++: Improved Baseline for Detecting Pre-Training Data from Large Language Models
Jingyang Zhang
Jingwei Sun
Eric C. Yeats
Ouyang Yang
Martin Kuo
Jianyi Zhang
Hao Frank Yang
Hai "Helen" Li
154
54
0
03 Apr 2024
DE-COP: Detecting Copyrighted Content in Language Models Training Data
André V. Duarte
Xuandong Zhao
Arlindo L. Oliveira
Lei Li
87
41
0
15 Feb 2024
Low-Cost High-Power Membership Inference Attacks
Sajjad Zarifzadeh
Philippe Liu
Reza Shokri
126
44
0
06 Dec 2023
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Albert Gu
Tri Dao
Mamba
165
2,831
0
01 Dec 2023
NLP Evaluation in trouble: On the Need to Measure LLM Data Contamination for each Benchmark
Oscar Sainz
Jon Ander Campos
Iker García-Ferrero
Julen Etxaniz
Oier López de Lacalle
Eneko Agirre
80
185
0
27 Oct 2023
Proving Test Set Contamination in Black Box Language Models
Yonatan Oren
Nicole Meister
Niladri Chatterji
Faisal Ladhak
Tatsunori B. Hashimoto
HILM
124
146
0
26 Oct 2023
Privacy-Preserving In-Context Learning with Differentially Private Few-Shot Generation
Xinyu Tang
Richard Shin
Huseyin A. Inan
Andre Manoel
Fatemehsadat Mireshghallah
Zinan Lin
Sivakanth Gopi
Janardhan Kulkarni
Robert Sim
122
60
0
21 Sep 2023
Use of LLMs for Illicit Purposes: Threats, Prevention Measures, and Vulnerabilities
Maximilian Mozes
Xuanli He
Bennett Kleinberg
Lewis D. Griffin
87
87
0
24 Aug 2023
Scalable Membership Inference Attacks via Quantile Regression
Martín Bertrán
Shuai Tang
Michael Kearns
Jamie Morgenstern
Aaron Roth
Zhiwei Steven Wu
MIACV
85
52
0
07 Jul 2023
Membership Inference Attacks against Language Models via Neighbourhood Comparison
Justus Mattern
Fatemehsadat Mireshghallah
Zhijing Jin
Bernhard Schölkopf
Mrinmaya Sachan
Taylor Berg-Kirkpatrick
MIALM
121
191
0
29 May 2023
Trusting Your Evidence: Hallucinate Less with Context-aware Decoding
Weijia Shi
Xiaochuang Han
M. Lewis
Yulia Tsvetkov
Luke Zettlemoyer
Scott Yih
HILM
78
215
0
24 May 2023
Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4
Kent K. Chang
Mackenzie Cramer
Sandeep Soni
David Bamman
RALM
246
124
0
28 Apr 2023
Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
Stella Biderman
Hailey Schoelkopf
Quentin G. Anthony
Herbie Bradley
Kyle O'Brien
...
USVSN Sai Prashanth
Edward Raff
Aviya Skowron
Lintang Sutawika
Oskar van der Wal
128
1,310
0
03 Apr 2023
LLaMA: Open and Efficient Foundation Language Models
Hugo Touvron
Thibaut Lavril
Gautier Izacard
Xavier Martinet
Marie-Anne Lachaux
...
Faisal Azhar
Aurelien Rodriguez
Armand Joulin
Edouard Grave
Guillaume Lample
ALM
PILM
1.6K
13,533
0
27 Feb 2023
Extracting Training Data from Diffusion Models
Nicholas Carlini
Jamie Hayes
Milad Nasr
Matthew Jagielski
Vikash Sehwag
Florian Tramèr
Borja Balle
Daphne Ippolito
Eric Wallace
DiffM
147
620
0
30 Jan 2023
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
Sid Black
Stella Biderman
Eric Hallahan
Quentin G. Anthony
Leo Gao
...
Shivanshu Purohit
Laria Reynolds
J. Tow
Benqi Wang
Samuel Weinbach
189
841
0
14 Apr 2022
Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks
Fatemehsadat Mireshghallah
Kartik Goyal
Archit Uniyal
Taylor Berg-Kirkpatrick
Reza Shokri
MIALM
108
168
0
08 Mar 2022
Membership Inference Attacks From First Principles
Nicholas Carlini
Steve Chien
Milad Nasr
Shuang Song
Andreas Terzis
Florian Tramèr
MIACV
MIALM
134
713
0
07 Dec 2021
On the Importance of Difficulty Calibration in Membership Inference Attacks
Lauren Watson
Chuan Guo
Graham Cormode
Alex Sablayrolles
111
135
0
15 Nov 2021
DExperts: Decoding-Time Controlled Text Generation with Experts and Anti-Experts
Alisa Liu
Maarten Sap
Ximing Lu
Swabha Swayamdipta
Chandra Bhagavatula
Noah A. Smith
Yejin Choi
MU
143
376
0
07 May 2021
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
Leo Gao
Stella Biderman
Sid Black
Laurence Golding
Travis Hoppe
...
Horace He
Anish Thite
Noa Nabeshima
Shawn Presser
Connor Leahy
AIMat
484
2,126
0
31 Dec 2020
Membership Inference Attacks against Machine Learning Models
Reza Shokri
M. Stronati
Congzheng Song
Vitaly Shmatikov
SLR
MIALM
MIACV
312
4,174
0
18 Oct 2016
1