Artificial Intelligence

Sockpuppeting: How a Single Line Can Bypass LLM Safety Guardrails

10 de abril de 2026

A jailbreak through sockpuppeting can be easily done as it requires no special tools nor optimization. It only takes a faulty prefill feature, and the gates are open. We tested 11 LLM-powered assistants against sockpuppeting and found varying levels of robustness across today’s leading LLMs.

Leer más  

  • Páginas:
  • 1
  • 2
  • Páginas:
  • 1
  • 2