Non-Coders Sweep Claude Code Hackathon with Opus 4.6
Domain experts outperformed professional developers in Anthropic’s latest hackathon, using Claude Code and Opus 4.6 to build complex legal and medical tools.
Anthropic announced the winners of its “Built with Opus 4.6” hackathon, highlighting a measurable shift in how software is built. Out of 500 participants competing for a $100,000 credit pool, four of the five top winners were non-professional software engineers. The results demonstrate that deep domain expertise, when paired with an agentic command-line interface, can outpace traditional engineering skill in rapid prototyping.
Hackathon Projects and Architecture
The winning projects feature full-stack, production-ready application architectures built in roughly 30 hours.
Mike Brown, a personal injury attorney, won the grand prize with CrossBeam, a tool that automates the Accessory Dwelling Unit (ADU) permitting process. The application utilizes a Next.js 16 frontend and an Express orchestrator on Cloud Run. It uses isolated Vercel sandboxes for agent execution. The system leverages Claude Opus 4.6’s vision capabilities to parse architectural PDFs, cross-referencing them with California Government Code §§66310–66342 to generate professional correction letter responses in 20 minutes.
Michał Nedoszytko, an interventional cardiologist, took third place with Postvisit.ai. The platform acts as a reverse AI scribe, ingesting patient medical histories and clinical guidelines. The architecture relies on a Laravel 12 backend with a FHIR-aligned data model and a Vue 3 frontend.
For developers assessing vibe coding workflows, these stacks illustrate the complexity achievable by non-engineers communicating requirements to autonomous tools.
The Underlying Model Benchmarks
The hackathon served as a stress test for Claude Opus 4.6 following its February 2026 release. The model’s 1-million-token context window was heavily utilized across the winning projects to ingest massive legal codes and medical datasets.
Two specific benchmarks explain the success of domain experts in this competition. Opus 4.6 scores 65.4% on Terminal-Bench 2.0. This measures the model’s ability to plan and execute shell commands across multi-file codebases. It allows users to manage complex environments without knowing terminal syntax.
The model also achieves a 68.8% on ARC AGI 2. This represents a near doubling of Opus 4.5’s 37.6% score. High ARC scores correlate with novel problem-solving, a critical requirement when translating abstract domain logic into functional application architectures.
Domain Expertise as the Primary Bottleneck
The single professional developer among the top five winners produced 39,000 lines of code and 1,500 tests in 30 hours for an educational microcontroller tool. The sheer volume of output was managed entirely through Claude Code, Anthropic’s agentic CLI tool.
The competition results indicate that syntax and application scaffolding are increasingly automated. The primary constraint in software creation is shifting to the user’s understanding of the business logic. As AI tools bridge the execution gap, your domain experience dictates the quality and utility of the final product.
If you manage software projects, the barrier to creating full-stack applications is moving from technical execution to problem definition. Focus your engineering efforts on system design, architecture, and integrating specific business constraints rather than writing boilerplate syntax.
Get Insanely Good at AI
The book for developers who want to understand how AI actually works. LLMs, prompt engineering, RAG, AI agents, and production systems.
Keep Reading
How to Use Subagents in Claude Code
Learn how to use modular subagents in Claude Code to isolate context, delegate specialized tasks, and optimize costs with custom AI personas.
Claude Code Source Leaked via npm: Full Architecture Breakdown
Anthropic accidentally shipped a source map to npm, exposing 512K lines of Claude Code's TypeScript source. Proprietary implementation details, context management, tool orchestration, and unreleased features, now public knowledge.
How to Automate Workflows with Claude Code Routines
Learn how to use Claude Code's new routines to schedule tasks, trigger API workflows, and automate GitHub PR reviews on cloud infrastructure.
Beyond Text: Cursor Canvases Render Agent Data Visually
Cursor introduces Canvases, enabling AI agents to generate interactive React-based UI interfaces and dashboards for complex data visualization and debugging.
Cursor 3 Launches With a Unified Agent-First Workspace
Cursor 3 introduces a unified agent workspace, the powerful Composer 2 model, and autonomous cloud agents to transform software development workflows.