Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2505.18773
Cited By
Strong Membership Inference Attacks on Massive Datasets and (Moderately) Large Language Models
24 May 2025
Jamie Hayes
Ilia Shumailov
Christopher A. Choquette-Choo
Matthew Jagielski
G. Kaissis
Katherine Lee
Milad Nasr
Sahra Ghalebikesabi
Niloofar Mireshghallah
Meenatchi Sundaram Mutu Selva Annamalai
Igor Shilov
Matthieu Meeus
Yves-Alexandre de Montjoye
Franziska Boenisch
Adam Dziedzic
A. Feder Cooper
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Strong Membership Inference Attacks on Massive Datasets and (Moderately) Large Language Models"
39 / 39 papers shown
Title
Extracting memorized pieces of (copyrighted) books from open-weight language models
A. Feder Cooper
Aaron Gokaslan
Amy Cyphert
Christopher De Sa
Mark A. Lemley
Daniel E. Ho
Percy Liang
38
2
0
18 May 2025
Privacy Auditing of Large Language Models
Ashwinee Panda
Xinyu Tang
Milad Nasr
Christopher A. Choquette-Choo
Prateek Mittal
PILM
128
10
0
09 Mar 2025
The Canary's Echo: Auditing Privacy Risks of LLM-Generated Synthetic Text
Matthieu Meeus
Lukas Wutschitz
Santiago Zanella Béguelin
Shruti Tople
Reza Shokri
166
1
0
19 Feb 2025
Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
A. Feder Cooper
Christopher A. Choquette-Choo
Miranda Bogen
Matthew Jagielski
Katja Filippova
...
Abigail Z. Jacobs
Andreas Terzis
Hanna M. Wallach
Nicolas Papernot
Katherine Lee
AILaw
MU
176
20
0
09 Dec 2024
Membership Inference Attacks Cannot Prove that a Model Was Trained On Your Data
Jie Zhang
Debeshee Das
Gautam Kamath
Florian Tramèr
MIALM
MIACV
296
27
1
29 Sep 2024
Context-Aware Membership Inference Attacks against Pre-trained Large Language Models
Hongyan Chang
Ali Shahin Shamsabadi
Kleomenis Katevas
Hamed Haddadi
Reza Shokri
MIALM
139
7
0
11 Sep 2024
Con-ReCall: Detecting Pre-training Data in LLMs via Contrastive Decoding
Cheng Wang
Yiwei Wang
Bryan Hooi
Yujun Cai
Nanyun Peng
Kai-Wei Chang
127
6
0
05 Sep 2024
Blind Baselines Beat Membership Inference Attacks for Foundation Models
Debeshee Das
Jie Zhang
Florian Tramèr
MIALM
166
39
1
23 Jun 2024
ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Roy Xie
Junlin Wang
Ruomin Huang
Minxing Zhang
Rong Ge
Jian Pei
Neil Zhenqiang Gong
Bhuwan Dhingra
MIALM
126
17
0
23 Jun 2024
LLM Dataset Inference: Did you train on my dataset?
Pratyush Maini
Hengrui Jia
Nicolas Papernot
Adam Dziedzic
MIALM
137
46
0
10 Jun 2024
MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Shengding Hu
Yuge Tu
Xu Han
Chaoqun He
Ganqu Cui
...
Chaochao Jia
Guoyang Zeng
Dahai Li
Zhiyuan Liu
Maosong Sun
MoE
111
345
0
09 Apr 2024
Proving membership in LLM pretraining data via data watermarks
Johnny Tian-Zheng Wei
Ryan Yixiang Wang
Robin Jia
WaLM
116
29
0
16 Feb 2024
DE-COP: Detecting Copyrighted Content in Language Models Training Data
André V. Duarte
Xuandong Zhao
Arlindo L. Oliveira
Lei Li
87
41
0
15 Feb 2024
Copyright Traps for Large Language Models
Matthieu Meeus
Igor Shilov
Manuel Faysse
Yves-Alexandre de Montjoye
105
22
0
14 Feb 2024
Do Membership Inference Attacks Work on Large Language Models?
Michael Duan
Anshuman Suri
Niloofar Mireshghallah
Sewon Min
Weijia Shi
Luke Zettlemoyer
Yulia Tsvetkov
Yejin Choi
David Evans
Hanna Hajishirzi
MIALM
130
100
0
12 Feb 2024
Beyond Chinchilla-Optimal: Accounting for Inference in Language Model Scaling Laws
Nikhil Sardana
Jacob P. Portes
Sasha Doubov
Jonathan Frankle
LRM
366
88
0
31 Dec 2023
Low-Cost High-Power Membership Inference Attacks
Sajjad Zarifzadeh
Philippe Liu
Reza Shokri
114
44
0
06 Dec 2023
Scalable Extraction of Training Data from (Production) Language Models
Milad Nasr
Nicholas Carlini
Jonathan Hayase
Matthew Jagielski
A. Feder Cooper
Daphne Ippolito
Christopher A. Choquette-Choo
Eric Wallace
Florian Tramèr
Katherine Lee
SILM
73
356
0
28 Nov 2023
Investigating Data Contamination in Modern Benchmarks for Large Language Models
Chunyuan Deng
Yilun Zhao
Xiangru Tang
Mark B. Gerstein
Arman Cohan
AAML
ELM
90
63
0
16 Nov 2023
Practical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt Calibration
Wenjie Fu
Huandong Wang
Chen Gao
Guanghua Liu
Yong Li
Tao Jiang
MIALM
77
49
0
10 Nov 2023
Proving Test Set Contamination in Black Box Language Models
Yonatan Oren
Nicole Meister
Niladri Chatterji
Faisal Ladhak
Tatsunori B. Hashimoto
HILM
117
146
0
26 Oct 2023
Detecting Pretraining Data from Large Language Models
Weijia Shi
Anirudh Ajith
Mengzhou Xia
Yangsibo Huang
Daogao Liu
Terra Blevins
Danqi Chen
Luke Zettlemoyer
MIALM
105
201
0
25 Oct 2023
Did the Neurons Read your Book? Document-level Membership Inference for Large Language Models
Matthieu Meeus
Shubham Jain
Marek Rei
Yves-Alexandre de Montjoye
MIALM
83
33
0
23 Oct 2023
User Inference Attacks on Large Language Models
Nikhil Kandpal
Krishna Pillutla
Alina Oprea
Peter Kairouz
Christopher A. Choquette-Choo
Zheng Xu
SILM
AAML
132
19
0
13 Oct 2023
Qwen Technical Report
Jinze Bai
Shuai Bai
Yunfei Chu
Zeyu Cui
Kai Dang
...
Zhenru Zhang
Chang Zhou
Jingren Zhou
Xiaohuan Zhou
Tianhang Zhu
OSLM
271
1,912
0
28 Sep 2023
Membership Inference Attacks against Language Models via Neighbourhood Comparison
Justus Mattern
Fatemehsadat Mireshghallah
Zhijing Jin
Bernhard Schölkopf
Mrinmaya Sachan
Taylor Berg-Kirkpatrick
MIALM
116
190
0
29 May 2023
LLaMA: Open and Efficient Foundation Language Models
Hugo Touvron
Thibaut Lavril
Gautier Izacard
Xavier Martinet
Marie-Anne Lachaux
...
Faisal Azhar
Aurelien Rodriguez
Armand Joulin
Edouard Grave
Guillaume Lample
ALM
PILM
1.6K
13,495
0
27 Feb 2023
Analyzing Leakage of Personally Identifiable Information in Language Models
Nils Lukas
A. Salem
Robert Sim
Shruti Tople
Lukas Wutschitz
Santiago Zanella Béguelin
PILM
155
235
0
01 Feb 2023
The Privacy Onion Effect: Memorization is Relative
Nicholas Carlini
Matthew Jagielski
Chiyuan Zhang
Nicolas Papernot
Andreas Terzis
Florian Tramèr
PILM
MIACV
136
110
0
21 Jun 2022
Quantifying Privacy Risks of Masked Language Models Using Membership Inference Attacks
Fatemehsadat Mireshghallah
Kartik Goyal
Archit Uniyal
Taylor Berg-Kirkpatrick
Reza Shokri
MIALM
89
166
0
08 Mar 2022
Quantifying Memorization Across Neural Language Models
Nicholas Carlini
Daphne Ippolito
Matthew Jagielski
Katherine Lee
Florian Tramèr
Chiyuan Zhang
PILM
127
631
0
15 Feb 2022
Membership Inference Attacks From First Principles
Nicholas Carlini
Steve Chien
Milad Nasr
Shuang Song
Andreas Terzis
Florian Tramèr
MIACV
MIALM
91
711
0
07 Dec 2021
Enhanced Membership Inference Attacks against Machine Learning Models
Jiayuan Ye
Aadyaa Maddi
S. K. Murakonda
Vincent Bindschaedler
Reza Shokri
MIALM
MIACV
105
256
0
18 Nov 2021
On the Importance of Difficulty Calibration in Membership Inference Attacks
Lauren Watson
Chuan Guo
Graham Cormode
Alex Sablayrolles
109
134
0
15 Nov 2021
Deduplicating Training Data Makes Language Models Better
Katherine Lee
Daphne Ippolito
A. Nystrom
Chiyuan Zhang
Douglas Eck
Chris Callison-Burch
Nicholas Carlini
SyDa
362
637
0
14 Jul 2021
Extracting Training Data from Large Language Models
Nicholas Carlini
Florian Tramèr
Eric Wallace
Matthew Jagielski
Ariel Herbert-Voss
...
Tom B. Brown
Basel Alomair
Ulfar Erlingsson
Alina Oprea
Colin Raffel
MLAU
SILM
528
1,958
0
14 Dec 2020
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Colin Raffel
Noam M. Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
AIMat
530
20,378
0
23 Oct 2019
White-box vs Black-box: Bayes Optimal Strategies for Membership Inference
Alexandre Sablayrolles
Matthijs Douze
Yann Ollivier
Cordelia Schmid
Hervé Jégou
MIACV
86
369
0
29 Aug 2019
Membership Inference Attacks against Machine Learning Models
Reza Shokri
M. Stronati
Congzheng Song
Vitaly Shmatikov
SLR
MIALM
MIACV
285
4,171
0
18 Oct 2016
1