Auditing Information Disclosure During LLM-Scale Gradient Descent Using Gradient Uniqueness
arXiv:2510.10902v2 Announce Type: replace-cross Abstract: Disclosing information via the publication of a machine learning model poses significant privacy risks. However, auditing this disclosure across every datapoint during the training of Large Language Models (LLMs) is computationally prohibitive. In this paper, we present Gradient Uniqueness (GNQ), a principled, attack-agnostic metric derived from an information-theoretic upper bound on the amount of information embedded in a model about individual training points via gradient descent. While naively computing GNQ requires forming and […]