Show HN: AI at Risk, a silly LLM benchmark
Summary
A developer created "AI at Risk," a playful benchmark where four AI agents with distinct personas compete in the board game Risk, using various language models. The new "cloaked" Horizon Alpha model has shown strong performance, outperforming others in the game. While not a rigorous evaluation, the project highlights the potential for creative, interactive AI benchmarks and offers insights into model behavior in complex, strategic environments.