R.R.: Unveiling LLM Training Privacy through Recollection and Ranking
FOS: Computer and information sciences
Computer Science - Computation and Language
Computation and Language (cs.CL)
DOI:
10.48550/arxiv.2502.12658
Publication Date:
2025-02-18
AUTHORS (8)
ABSTRACT
Large Language Models (LLMs) pose significant privacy risks, potentially leaking training data due to implicit memorization. Existing attacks primarily focus on membership inference (MIAs) or extraction attacks, but reconstructing specific personally identifiable information (PII) in LLM's remains challenging. In this paper, we propose R.R. (Recollect and Rank), a novel two-step stealing attack that enables attackers reconstruct PII entities from scrubbed where the have been masked. first stage, introduce prompt paradigm named recollection, which instructs LLM repeat masked text fill masks. Then can use identifiers extract recollected candidates. second design new criterion score each candidate rank them. Motivated by inference, leverage reference model as calibration our criterion. Experiments across three popular datasets demonstrate achieves better identical performance compared baselines. These results highlight vulnerability of LLMs leakage even when has scrubbed. We release replicate package at link.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....