Is Multilingual LLM Watermarking Truly Multilingual? A Simple Back-Translation Solution Paper β’ 2510.18019 β’ Published 17 days ago β’ 17
Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models Paper β’ 2411.00154 β’ Published Oct 31, 2024 β’ 1
TRAP: Targeted Random Adversarial Prompt Honeypot for Black-Box Identification Paper β’ 2402.12991 β’ Published Feb 20, 2024 β’ 1
Calibrating Large Language Models Using Their Generations Only Paper β’ 2403.05973 β’ Published Mar 9, 2024 β’ 1
ProPILE: Probing Privacy Leakage in Large Language Models Paper β’ 2307.01881 β’ Published Jul 4, 2023 β’ 2
DISCO: Diversifying Sample Condensation for Efficient Model Evaluation Paper β’ 2510.07959 β’ Published 28 days ago β’ 14
Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers Paper β’ 2506.15674 β’ Published Jun 18 β’ 2
MIA-Pile Collection Samples used for the NAACL 2025 Findings paper: "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models." β’ 23 items β’ Updated Feb 3 β’ 1
π Scaling MIA Data & Results Collection NAACL 2025 Findings "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models" https://arxiv.org/abs/2411.00154 β’ 22 items β’ Updated Jun 5 β’ 2
π Papers Collection List of research articles of Parameter Lab β’ 8 items β’ Updated 15 days ago β’ 1
Model with Circuit Breakers Collection SoTA models with circuit breakers inserted. Top safety performance without losing capabilities. β’ 3 items β’ Updated Oct 25, 2024 β’ 5
π Apricot Models Collection Fine-tuned models for black-box LLM calibration, trained for "Apricot: Calibrating Large Language Models Using Their Generations Only" (ACL 2024) β’ 9 items β’ Updated Nov 20, 2024 β’ 3