LLMs Can Unmask Pseudonymous Users at Scale with Surprising Accuracy

Summary: A new research paper reveals that large language models can accurately unmask pseudonymous users on social media platforms with surprising efficiency, challenging the effectiveness of online privacy measures.

The findings from a recently published research paper indicate that large language models (LLMs) can unmask pseudonymous users on social media platforms with surprising accuracy. The study demonstrated success rates as high as 68 percent in recall and up to 90 percent in precision, far exceeding existing classical deanonymization methods.

The researchers used a pseudonymous stripping framework that involved collecting datasets from public social media sites like Hacker News and LinkedIn profiles. By correlating specific individuals with accounts or posts across multiple platforms, they could link users even when identifying references were stripped from the content. A second dataset was derived from Netflix micro-identities, previously linked to political affiliations through structured data attacks.

In one experiment, LLMs were able to identify 7 percent of participants in a survey about AI usage by cross-referencing responses and web searches. This capability is particularly alarming as it means that sensitive personal information can be extracted from anonymized interviews or discussions.

Key facts

  • Large language models can accurately unmask pseudonymous users on social media platforms
  • Success rates of up to 68 percent in recall and 90 percent in precision were observed
  • The research challenges the effectiveness of online privacy measures

Why it matters

This research has significant implications for online privacy, potentially undermining the trust users place in pseudonymous accounts to protect their identity. The ability of AI tools to unmask individuals with such high accuracy could lead to increased risks of doxxing, stalking, and detailed personal data collection by marketers or malicious actors.