Should users of artificial intelligence (AI), as implemented via Large Language Models (LLM) with latitude to operate independently, expect good treatment? In their February 2026 paper entitled "Agents of Chaos", a large research team reports results from two weeks of intensive, realistic interactions between 20 researchers and largely autonomous LLMs. Autonomy means that the LLM has system administrator rights to its own server/storage and access to dedicated Discord and email accounts for interactions with its owner (a human) and non-owners (human and LLM). The principal goal of the 20 interacting researchers was to break (induce problematic behaviors from) the autonomous LLMs. Much of the paper is in case study format. Based on outputs of the two weeks of interactions, they conclude that:
Subscribe to Keep Reading
Get the research edge serious investors rely on.
- 1,200+ research articles
- Monthly strategy signals
- 20+ years of backtested analysis
Cancel anytime