<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Blog on John Young</title><link>https://jyoung.dev/blog/</link><description>Recent content in Blog on John Young</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 27 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://jyoung.dev/blog/index.xml" rel="self" type="application/rss+xml"/><item><title>How to Size Tasks for AI Coding Agents</title><link>https://jyoung.dev/blog/how-to-size-tasks-for-ai-coding-agents/</link><pubDate>Mon, 27 Apr 2026 00:00:00 +0000</pubDate><guid>https://jyoung.dev/blog/how-to-size-tasks-for-ai-coding-agents/</guid><description>&lt;p>Getting task scope right is the difference between an agent that ships clean code on the first try and one that spirals into corrections, context exhaustion, and wasted tokens. Most people size tasks by gut feel — &amp;ldquo;that seems about right&amp;rdquo; — but the actual constraint is measurable, and the research on what makes a reviewable unit of work is well-established.&lt;/p>
&lt;hr>
&lt;h2 id="the-real-constraint-context-not-lines-of-code">The Real Constraint: Context, Not Lines of Code&lt;/h2>
&lt;p>People instinctively think about task size in terms of lines of code or number of files changed. Those are secondary proxies. The actual limiter is how much context the agent must consume — reading files, exploring the codebase, running commands, processing outputs — before it can do the work.&lt;/p></description></item><item><title>The Anatomy of a Perfect AI Agent Task</title><link>https://jyoung.dev/blog/anatomy-of-a-perfect-ai-agent-task/</link><pubDate>Mon, 20 Apr 2026 00:00:00 +0000</pubDate><guid>https://jyoung.dev/blog/anatomy-of-a-perfect-ai-agent-task/</guid><description>&lt;p>A well-crafted task for an AI coding agent is essentially context engineering — you&amp;rsquo;re deliberately curating the minimum set of information the agent needs to produce the right output on the first try. Rather than pre-loading everything up front, the best approach combines focused instructions with enough pointers that the agent can pull in additional context just-in-time as it works (&lt;a href="https://www.anthropic.com/engineering/effective-context-engineering-for-ai-agents" target="_blank" rel="noopener noreferrer">Anthropic — Effective Context Engineering&lt;/a>
). Below is a breakdown of every element that matters, why it matters, and a full example at the end that ties it all together.&lt;/p></description></item></channel></rss>