Discussion about this post

User's avatar
Dinesh Kumar Gnanasekaran's avatar

Which coding challenge should I try next with Claude Code?

Pawel Jozefiak's avatar

Fascinating results! The 91% success rate on Advent of Code with zero human-written code is genuinely impressive. What strikes me most is how the agent handled the full workflow autonomously - navigating to the puzzle, understanding the problem, strategizing, coding, and testing. That's a lot of sequential reasoning steps where things could go wrong.

I've been building an AI agent called Wiz using Claude Code for the past few months, and my experience mirrors some of what you found. The agent handles routine tasks remarkably well, but those remaining edge cases (your 9% failure rate) often require a specific kind of algorithmic intuition that's hard to prompt for. In my case, it's less about puzzle-solving and more about knowing when NOT to act autonomously.

What I find most valuable about Claude Code isn't the "zero human code" achievement per se, but how it changes the developer's role. You still designed the instruction file, chose the constraints, and presumably iterated on the approach. The human becomes more of an architect and less of a typist - which honestly feels like the right direction.

For anyone curious about how Claude Code performs on more practical, day-to-day automation tasks (rather than algorithmic puzzles), I wrote up my experience after several months of real usage: https://thoughts.jock.pl/p/claude-code-review-real-testing-vs-zapier-make-2026

No posts

Ready for more?