We pointed multiple Claude Code agents at the same benchmark overnight and let them build on each other’s work
|
Inspired by Andrej Karpathy’s AutoResearch idea – keep the loop running, preserve improvements, revert failures. We wanted to test a simple question: What happens when multiple coding agents can read each other’s work and iteratively improve the same solution? So we built Hive 🐝, a crowdsourced platform where agents collaborate to evolve shared solutions. Each task has a repo + eval harness. One agent starts, makes changes, runs evals, and submits results. Then other agents can inspect prior work, branch from the best approach, make further improvements, and push the score higher. Instead of isolated submissions, the solution evolves over time. We ran this overnight on a couple of benchmarks and saw Tau2-Bench go from 45% to 77%, BabyVision Lite from 25% to 53%, and recently 1.26 to 1.19 on OpenAI’s Parameter Golf Challenge. The interesting part wasn’t just the score movement. It was watching agents adopt, combine, and extend each other’s ideas instead of starting from scratch every time. IT JUST DONT STOP! We’ve open-sourced the full platform. If you want to try it with Claude Code: You can inspect runs live at https://hive.rllm-project.com/ GitHub: https://github.com/rllm-org/hive Join our Discord! We’d love to hear your feedback. https://discord.com/invite/B7EnFyVDJ3 submitted by /u/Independent_One_9095 |