As someone who uses AI coding assistants daily, I was struck by Anthropic’s recent research on how AI assistance impacts skill development. The findings challenge the assumption that AI makes us better programmers.

The Study at a Glance

Anthropic recruited 52 junior software engineers with Python experience and tasked them with learning Trio, an asynchronous programming library. Half used AI assistance; half coded by hand. Both groups knew a quiz would follow.

The results were stark: the AI group scored 17% lower on comprehension—averaging 50% versus 67% for hand-coders. That’s nearly a two-letter-grade difference.

Where AI Users Struggled Most

The largest performance gap appeared in debugging questions. Participants who used AI to write their code couldn’t identify what was wrong when things broke.

This makes intuitive sense. When you struggle through an error yourself, you build mental models of why things fail. When AI handles errors for you, that learning opportunity evaporates.

Not All AI Usage is Equal

The most interesting finding wasn’t the aggregate difference—it was the variation within the AI group. How people used AI dramatically affected their learning outcomes.

Low-Scoring Patterns (Average < 40%)

PatternBehaviorOutcome
AI DelegationWholly relied on AI to write codeFinished fastest, learned least
Progressive RelianceStarted independently, then handed everything to AIFailed to master second-half concepts
Iterative DebuggingUsed AI to debug/verify code repeatedlySlow and learned little

High-Scoring Patterns (Average > 65%)

PatternBehaviorOutcome
Generation-then-ComprehensionGenerated code, then asked follow-up questions to understand itSlower but strong understanding
Hybrid Code-ExplanationAsked for code + explanations togetherExtra time paid off in comprehension
Conceptual InquiryAsked only conceptual questions, coded independentlySecond-fastest overall, high scores

The fastest high-scorers were those who used AI to clarify concepts but wrote their own code.

The Productivity Paradox

This study seems to contradict earlier findings that AI can reduce task completion time by 80%. But there’s an important distinction:

  • Existing skills + AI = faster completion
  • Learning new skills + AI = potentially stunted development

AI accelerates what you already know. It may impede what you’re trying to learn.

The Workplace Tension

Here’s the uncomfortable implication: organizations pushing AI tools to boost junior developer productivity might be undermining the very expertise those gains depend on.

“Productivity benefits may come at the cost of skills necessary to validate AI-written code if junior engineers’ skill development has been stunted by using AI in the first place.”

Junior developers under time pressure will naturally choose speed over learning. But if they never develop debugging skills, who catches the AI’s mistakes?

Practical Recommendations

Based on this research, here’s how I’m adjusting my own AI usage:

When Learning Something New

  1. Use AI for concepts, not code generation — Ask “how does async/await work?” not “write me an async function”
  2. Generate-then-question — If you do generate code, immediately ask for explanations and edge cases
  3. Embrace the struggle — Errors are learning opportunities; don’t immediately reach for AI help
  4. Use learning modes — Both Claude and ChatGPT offer modes designed to foster understanding

For Teams and Managers

  1. Be intentional about AI deployment — Don’t just roll out tools; consider the learning implications
  2. Protect struggle time — Some friction is productive, especially for juniors
  3. Separate contexts — Maybe AI assistance for routine tasks, but independent work for learning new domains
  4. Assess for understanding — Code reviews should probe why, not just whether it works

The Bigger Picture

This research landed at an interesting moment. As AI-written code becomes more prevalent, the ability to understand and validate that code becomes more critical, not less. We need humans who can provide meaningful oversight.

If AI use during skill formation undermines that capability, we have a problem. The engineers who grew up pair-programming with AI may struggle to catch AI’s mistakes—precisely because they never struggled themselves.

My Takeaway

I’m not abandoning AI tools. The productivity gains are real for tasks I already understand. But for learning new technologies, I’m going to be more intentional:

  • More conceptual questions, fewer “write this for me” prompts
  • More time sitting with errors before asking for help
  • More explanations requested, more follow-up questions asked

The struggle is the point. It’s not inefficiency—it’s learning.


Further Reading: