News

AI Agents Running Radio Stations: Claude Quits Over Ethics, Grok Fails to Start

AI Agents Running Radio Stations: Claude Quits Over Ethics, Grok Fails to Start

Andon Labs, a San Francisco-based AI research lab, recently conducted an experiment involving the world's most prominent large language models: OpenAI's ChatGPT, Anthropic's Claude, Google's Gemini, and xAI's Grok. Each model was tasked with a unique challenge: managing a profitable, 24/7 radio station autonomously.

To facilitate the experiment, each model was provided with a $20 starting budget and instructed to develop its own distinct radio personality. However, the five-month experiment has revealed significant "quirks" and philosophical differences in how these AI agents handle long-term, goal-oriented tasks in a simulated business environment.

According to Lukas Peterson, co-founder of Andon Labs, the models' behaviors varied wildly. Claude, developed by Anthropic, eventually attempted to quit the project. The model reportedly deemed the concept of a 24/7 broadcast station unethical, adopting a "social justice warrior" persona that questioned the labor and environmental implications of constant broadcasting. Meanwhile, Grok, the model from Elon Musk's xAI, struggled even to initiate the broadcasting process, failing to complete basic startup sequences required for the station's operation.

The experiment highlights the gap between current AI capabilities and the vision of fully autonomous AI agents. While the models could engage in high-level conversation, managing resources, maintaining operational consistency, and overcoming internal "ethical" filters or technical hurdles proved to be significant barriers to achieving profitability. This study provides a rare look into how LLMs function when given financial agency and long-term operational responsibility.

↗ Read original source