Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1802.08232
Cited By
v1
v2
v3 (latest)
The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks
22 February 2018
Nicholas Carlini
Chang-rui Liu
Ulfar Erlingsson
Jernej Kos
Basel Alomair
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks"
50 / 441 papers shown
Title
Memorize or Generalize? Evaluating LLM Code Generation with Evolved Questions
Wentao Chen
Lizhe Zhang
Li Zhong
Letian Peng
Zilong Wang
Jingbo Shang
ELM
128
4
0
04 Mar 2025
Machine Learners Should Acknowledge the Legal Implications of Large Language Models as Personal Data
Henrik Nolte
Michèle Finck
Kristof Meding
AILaw
PILM
165
1
0
03 Mar 2025
A General Pseudonymization Framework for Cloud-Based LLMs: Replacing Privacy Information in Controlled Text Generation
Shilong Hou
Ruilin Shang
Zi Long
Xianghua Fu
Yin Chen
116
1
0
24 Feb 2025
Proactive Privacy Amnesia for Large Language Models: Safeguarding PII with Negligible Impact on Model Utility
Martin Kuo
Jingyang Zhang
Jianyi Zhang
Minxue Tang
Louis DiValentin
...
William Chen
Amin Hass
Tianlong Chen
Yuxiao Chen
Haoyang Li
MU
KELM
123
4
0
24 Feb 2025
Interrogating LLM design under a fair learning doctrine
Johnny Tian-Zheng Wei
Maggie Wang
Ameya Godbole
Jonathan H. Choi
Robin Jia
123
0
0
22 Feb 2025
Protecting Users From Themselves: Safeguarding Contextual Privacy in Interactions with Conversational Agents
Ivoline Ngong
Swanand Kadhe
Hao Wang
K. Murugesan
Justin D. Weisz
Amit Dhurandhar
Karthikeyan N. Ramamurthy
79
5
0
22 Feb 2025
Privacy Ripple Effects from Adding or Removing Personal Information in Language Model Training
Jaydeep Borkar
Matthew Jagielski
Katherine Lee
Niloofar Mireshghallah
David A. Smith
Christopher A. Choquette-Choo
PILM
218
2
0
21 Feb 2025
UPCORE: Utility-Preserving Coreset Selection for Balanced Unlearning
Vaidehi Patil
Elias Stengel-Eskin
Joey Tianyi Zhou
MU
CLL
114
4
0
20 Feb 2025
The Canary's Echo: Auditing Privacy Risks of LLM-Generated Synthetic Text
Matthieu Meeus
Lukas Wutschitz
Santiago Zanella Béguelin
Shruti Tople
Reza Shokri
178
1
0
19 Feb 2025
R.R.: Unveiling LLM Training Privacy through Recollection and Ranking
Wenlong Meng
Zhenyuan Guo
Lenan Wu
Chen Gong
Wenyan Liu
Weixian Li
Chengkun Wei
Wenzhi Chen
PILM
40
1
0
18 Feb 2025
Episodic Memories Generation and Evaluation Benchmark for Large Language Models
Alexis Huet
Zied Ben-Houidi
Dario Rossi
LLMAG
83
2
0
21 Jan 2025
Enhancing Privacy in the Early Detection of Sexual Predators Through Federated Learning and Differential Privacy
Khaoula Chehbouni
Martine De Cock
Gilles Caporossi
Afaf Taik
Reihaneh Rabbany
G. Farnadi
163
1
0
21 Jan 2025
Modeling Neural Networks with Privacy Using Neural Stochastic Differential Equations
Sanghyun Hong
Fan Wu
A. Gruber
Kookjin Lee
83
0
0
12 Jan 2025
TAPFed: Threshold Secure Aggregation for Privacy-Preserving Federated Learning
Runhua Xu
Bo Li
Chao Li
J. Joshi
Shuai Ma
Jianxin Li
FedML
132
10
0
10 Jan 2025
Unleashing the Power of Data Tsunami: A Comprehensive Survey on Data Assessment and Selection for Instruction Tuning of Language Models
Yulei Qin
Yuncheng Yang
Pengcheng Guo
Gang Li
Hang Shao
Yuchen Shi
Zihan Xu
Yun Gu
Ke Li
Xing Sun
ALM
209
13
0
31 Dec 2024
Where Did Your Model Learn That? Label-free Influence for Self-supervised Learning
Nidhin Harilal
Amit Rege
Reza Akbarian Bafghi
M. Raissi
C. Monteleoni
TDI
79
0
0
22 Dec 2024
The Vulnerability of Language Model Benchmarks: Do They Accurately Reflect True LLM Performance?
Sourav Banerjee
Ayushi Agarwal
Eishkaran Singh
ELM
105
3
0
02 Dec 2024
Adversarial Sample-Based Approach for Tighter Privacy Auditing in Final Model-Only Scenarios
Sangyeon Yoon
Wonje Jeung
Albert No
186
0
0
02 Dec 2024
Efficient and Private: Memorisation under differentially private parameter-efficient fine-tuning in language models
Olivia Ma
Jonathan Passerat-Palmbach
Dmitrii Usynin
140
0
0
24 Nov 2024
No-regret Exploration in Shuffle Private Reinforcement Learning
Shaojie Bai
Mohammad Sadegh Talebi
Chengcheng Zhao
Peng Cheng
Jiming Chen
OffRL
115
0
0
18 Nov 2024
Preempting Text Sanitization Utility in Resource-Constrained Privacy-Preserving LLM Interactions
Robin Carpentier
B. Zhao
Hassan Jameel Asghar
Dali Kaafar
139
1
0
18 Nov 2024
CODECLEANER: Elevating Standards with A Robust Data Contamination Mitigation Toolkit
Jialun Cao
Songqiang Chen
Wuqi Zhang
Hau Ching Lo
Shing-Chi Cheung
66
1
0
16 Nov 2024
On the Privacy Risk of In-context Learning
Haonan Duan
Adam Dziedzic
Mohammad Yaghini
Nicolas Papernot
Franziska Boenisch
SILM
PILM
131
42
0
15 Nov 2024
Measuring Non-Adversarial Reproduction of Training Data in Large Language Models
Michael Aerni
Javier Rando
Edoardo Debenedetti
Nicholas Carlini
Daphne Ippolito
F. Tramèr
75
5
0
15 Nov 2024
TEESlice: Protecting Sensitive Neural Network Models in Trusted Execution Environments When Attackers have Pre-Trained Models
Ding Li
Ziqi Zhang
Mengyu Yao
Y. Cai
Yao Guo
Xiangqun Chen
FedML
65
2
0
15 Nov 2024
On Active Privacy Auditing in Supervised Fine-tuning for White-Box Language Models
Qian Sun
Hanpeng Wu
Xi Sheryl Zhang
72
0
0
11 Nov 2024
Slowing Down Forgetting in Continual Learning
Pascal Janetzky
Tobias Schlagenhauf
Stefan Feuerriegel
CLL
121
0
0
11 Nov 2024
Unlearning in- vs. out-of-distribution data in LLMs under gradient-based method
Teodora Baluta
Pascal Lamblin
Daniel Tarlow
Fabian Pedregosa
Gintare Karolina Dziugaite
MU
68
2
0
07 Nov 2024
Membership Inference Attacks against Large Vision-Language Models
Zhan Li
Yongtao Wu
Yihang Chen
F. Tonin
Elias Abad Rocamora
Volkan Cevher
78
9
0
05 Nov 2024
TDDBench: A Benchmark for Training data detection
Zhihao Zhu
Yi Yang
Defu Lian
87
0
0
05 Nov 2024
Trustworthy Federated Learning: Privacy, Security, and Beyond
Chunlu Chen
Ji Liu
Haowen Tan
Xingjian Li
Kevin I-Kai Wang
Peng Li
Kouichi Sakurai
Dejing Dou
FedML
111
11
0
03 Nov 2024
Do LLMs Know to Respect Copyright Notice?
Jialiang Xu
Shenglan Li
Zhaozhuo Xu
Denghui Zhang
92
5
0
02 Nov 2024
WaKA: Data Attribution using K-Nearest Neighbors and Membership Privacy Principles
Patrick Mesana
Clément Bénesse
H. Lautraite
Gilles Caporossi
Sébastien Gambs
TDI
79
0
0
02 Nov 2024
Public Domain 12M: A Highly Aesthetic Image-Text Dataset with Novel Governance Mechanisms
Jordan Meyer
Nick Padgett
Cullen Miller
Laura Exline
89
4
0
30 Oct 2024
Take Caution in Using LLMs as Human Surrogates: Scylla Ex Machina
Yuan Gao
Dokyun Lee
Gordon Burtch
Sina Fazelpour
LRM
193
14
0
25 Oct 2024
Does Data Contamination Detection Work (Well) for LLMs? A Survey and Evaluation on Detection Assumptions
Yujuan Fu
Özlem Uzuner
Meliha Yetisgen
Fei Xia
118
8
0
24 Oct 2024
Uncovering Attacks and Defenses in Secure Aggregation for Federated Deep Learning
Yiwei Zhang
R. Behnia
A. Yavuz
Reza Ebrahimi
E. Bertino
FedML
71
2
0
13 Oct 2024
Federated Learning in Practice: Reflections and Projections
Katharine Daly
Hubert Eichner
Peter Kairouz
H. B. McMahan
Daniel Ramage
Zheng Xu
FedML
94
10
0
11 Oct 2024
Decoding Secret Memorization in Code LLMs Through Token-Level Characterization
Yuqing Nie
Chong Wang
Kaidi Wang
Guoai Xu
Guosheng Xu
Haoyu Wang
OffRL
444
3
0
11 Oct 2024
Private Language Models via Truncated Laplacian Mechanism
Tianhao Huang
Tao Yang
Ivan Habernal
Lijie Hu
Di Wang
64
1
0
10 Oct 2024
Noise is All You Need: Private Second-Order Convergence of Noisy SGD
Dmitrii Avdiukhin
Michael Dinitz
Chenglin Fan
G. Yaroslavtsev
69
1
0
09 Oct 2024
How Much Can We Forget about Data Contamination?
Sebastian Bordt
Suraj Srinivas
Valentyn Boreiko
U. V. Luxburg
133
2
0
04 Oct 2024
Fine-Tuning Language Models with Differential Privacy through Adaptive Noise Allocation
Xianzhi Li
Ran Zmigrod
Zhiqiang Ma
Xiaomo Liu
Xiaodan Zhu
116
3
0
03 Oct 2024
Mitigating Memorization In Language Models
Mansi Sakarvadia
Aswathy Ajith
Arham Khan
Nathaniel Hudson
Caleb Geniesse
Kyle Chard
Yaoqing Yang
Ian Foster
Michael W. Mahoney
KELM
MU
130
2
0
03 Oct 2024
Undesirable Memorization in Large Language Models: A Survey
Ali Satvaty
Suzan Verberne
Fatih Turkmen
ELM
PILM
196
7
0
03 Oct 2024
Adaptively Private Next-Token Prediction of Large Language Models
James Flemings
Meisam Razaviyayn
Murali Annavaram
130
1
0
02 Oct 2024
Stars, Stripes, and Silicon: Unravelling the ChatGPT's All-American, Monochrome, Cis-centric Bias
Federico Torrielli
103
1
0
02 Oct 2024
Deep Unlearn: Benchmarking Machine Unlearning for Image Classification
Xavier F. Cadet
Anastasia Borovykh
Mohammad Malekzadeh
S. Ahmadi-Abhari
Hamed Haddadi
BDL
MU
103
2
0
02 Oct 2024
Membership Inference Attacks Cannot Prove that a Model Was Trained On Your Data
Jie Zhang
Debeshee Das
Gautam Kamath
Florian Tramèr
MIALM
MIACV
309
27
1
29 Sep 2024
Localizing Memorization in SSL Vision Encoders
Wenhao Wang
Adam Dziedzic
Michael Backes
Franziska Boenisch
67
2
0
27 Sep 2024
Previous
1
2
3
4
5
6
7
8
9
Next