this post was submitted on 27 Feb 2026
56 points (98.3% liked)

science

25523 readers
608 users here now

A community to post scientific articles, news, and civil discussion.

dart board;; science bs

rule #1: be kind

founded 2 years ago
MODERATORS
 

We show that large language models can be used to perform at-scale deanonymization. With full Internet access, our agent can re-identify Hacker News users and Anthropic Interviewer participants at high precision, given pseudonymous online profiles and conversations alone, matching what would take hours for a dedicated human investigator. We then design attacks for the closed-world setting. Given two databases of pseudonymous individuals, each containing unstructured text written by or about that individual, we implement a scalable attack pipeline that uses LLMs to: (1) extract identity-relevant features, (2) search for candidate matches via semantic embeddings, and (3) reason over top candidates to verify matches and reduce false positives. Compared to classical deanonymization work (e.g., on the Netflix prize) that required structured data, our approach works directly on raw user content across arbitrary platforms. We construct three datasets with known ground-truth data to evaluate our attacks. The first links Hacker News to LinkedIn profiles, using cross-platform references that appear in the profiles. Our second dataset matches users across Reddit movie discussion communities; and the third splits a single user's Reddit history in time to create two pseudonymous profiles to be matched. In each setting, LLM-based methods substantially outperform classical baselines, achieving up to 68% recall at 90% precision compared to near 0% for the best non-LLM method. Our results show that the practical obscurity protecting pseudonymous users online no longer holds and that threat models for online privacy need to be reconsidered.

top 8 comments
sorted by: hot top controversial new old
[–] WesternInfidels@feddit.online 1 points 6 hours ago

I disguise my internet writings by doing them left handed.

[–] eleijeep@piefed.social 5 points 9 hours ago (1 children)

The false positives are going to be tremendously dangerous to the falsely accused, as there is no way to either prove or disprove the charge. It's just a case of "computer says you're 99% likely to be the same person."

[–] WesternInfidels@feddit.online 1 points 6 hours ago

In a sane world, the computer's calculation would be used to guide a real investigation, which would find real evidence. The statistics wouldn't be used as proof of anything.

[–] lexiw@lemmy.world 4 points 13 hours ago

None of what this does is new, the issue is that’s being commodified: in a short amount of time these techniques will be available to non experts in the field.

[–] Wigners_friend@piefed.social 12 points 16 hours ago* (last edited 16 hours ago) (1 children)

So they found something LLMs can actually do, and predictably it's incredibly evil.

[–] 4am@lemmy.zip 5 points 15 hours ago

This is why they are building huge data centers for something nobody actually wants.

[–] theunknownmuncher@lemmy.world 10 points 16 hours ago (1 children)

That’s why I always run my text through an LLM for refinement before posting it. 😉

[–] WesternInfidels@feddit.online 1 points 6 hours ago

"ChatGPT, please rewrite this inflammatory post full of libelous accusations in the style of my worst enemy."