Here we share Nicholas Carlini – a Research Scientist at Anthropic – speaking at the [un]prompted 2026 conference on: Black-hat LLMs and how they uncover vulnerabilities (vulns).
Large language models are now capable of automating attacks that were previously only possible by human adversaries. In this talk, I discuss several ways that adversaries could mis-use curr...