LLMs can unmask pseudonymous users at scale with surprising accuracy


Recall at various precision thresholds.

Recall at various precision thresholds.

In a third experiment, the researchers took 5,000 users from the Netflix dataset and added another 5,000 “distraction” identities of people not in the results. They then added to the list of 10,000 candidate profiles 5,000 query distractors comprising users who appear only in a query set, with no true match in the candidate pool.

Compared to a classical baseline that mimics the Netflix Prize attack to LLM deanonymization, the latter far outperformed the former.

Screenshot

The researchers wrote:

(a) The precision of classical attacks drops very fast, explaining its low recall. In contrast, the precision of LLM-based attacks decays more gracefully as the attacker makes more guesses. (b) The classical attack almost fails completely even at moderately low precision. In contrast, even the simplest LLM attack (Search) achieves non-trivial recall at low precision, and extending it with Reason and Calibrate steps doubles Recall @99% Precision.

The results show that LLMs, while still prone to false positives and other weaknesses, are quickly outstripping more traditional, resource-intensive methods for identifying users online.

The researchers went on to propose mitigations, including platforms enforcing rate limits on API access to user data, detecting automated scraping, and restricting bulk data exports. LLM providers could also monitor for the misuse of their models in deanonymization attacks and build guardrails that make models refuse deanonymization requests.

Of course, another option is for people to dramatically curb their use of social media, or at a minimum, regularly delete posts after a set time threshold.

If LLMs’ success in deanonymizing people improves, the researchers warn, governments could use the techniques to unmask online critics, corporations can assemble customer profiles for “hyper-targeted advertising,” and attackers could build profiles of targets at scale to launch highly personalized social engineering scams.

“Recent advances in LLM capabilities have made it clear that there is an urgent need to rethink various aspects of computer security in the wake of LLM-driven offensive cyber capabilities, the researchers warned. “Our work shows that the same is likely true for privacy as well.”



Source link

  • Related Posts

    X begins testing standalone X Chat app on iOS

    Social network X is bringing its private messaging service, dubbed X Chat, to a standalone app. The company on Monday said the initial beta of the X Chat app was…

    Five of the most interesting upcoming indie games | Games

    These days, it’s easy to fall into the trap of thinking that every new indie game is either a co-op extraction shooter or a roguelike deck-builder – fortunately that’s not…

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    We Spoke With Lindsay Dahl On Clean Living, Labels, And Laws

    We Spoke With Lindsay Dahl On Clean Living, Labels, And Laws

    Feeling deflated

    7 Of The World’s Most Notable New Airline Routes This Week

    7 Of The World’s Most Notable New Airline Routes This Week

    Marathon post-launch content includes free seasonal updates, new Runner shell, and more – PlayStation.Blog

    Marathon post-launch content includes free seasonal updates, new Runner shell, and more – PlayStation.Blog

    Gas prices jump over 10 cents in a day, set to rise ‘very quickly’ as oil surges amid Iran war

    Gas prices jump over 10 cents in a day, set to rise ‘very quickly’ as oil surges amid Iran war

    Potential new Iranian power broker has extensive ties to Canada