Admios Initial Letter

The Agent Wrangler

There's a version of "AI-assisted development" that most engineering teams are doing right now: you write code, Claude suggests the next line, you accept or reject. It's useful. It's also not what we're talking about here.

Over the past several months, we've been embedded with a client team, working as a natural extension of their engineering org, that took a different position. One that's less about AI as autocomplete and more about AI as a co-worker that drafts, tests, and opens pull requests while you review and guide.

The client's direction was unambiguous: they would rather pay for tokens than bring on additional engineers. The team tracks token usage and treats high consumption as a proxy for productivity, not a cost concern.

What followed was instructive, not because everything went perfectly, but because of what the team had to build, change, and rethink before that model actually worked.

This is what we mean when we say AI doesn't replace great engineering, it amplifies it. The capability is real, but only when the engineering fundamentals are in place.

---

Your Repo Probably Isn't Ready

The biggest misconception about agentic development is that you point Claude at your codebase and it starts doing useful work. Sometimes it does. More often, you discover your repo is optimized for humans and hostile to agents.

Humans tolerate 1,500-line files because they have context. They remember what `utils.js` is for. They know which functions are deprecated. Claude doesn't have that. It works from what it can read in a single context window, and a sprawling file is a context problem.

The team we worked with built a set of conventions — call it an **agentic repo checklist** — that addressed this head-on:

File size discipline. Every file stays under 300 lines. This is stricter than most style guides, and the rationale is specifically about LLM context: a focused, well-named file is far easier for an agent to reason over than a monolith. The discipline forces decomposition that turns out to be good engineering regardless of AI.

Dockerized commands. Claude runs commands. If those commands depend on your local machine state — a specific environment variable, a path that exists on your laptop but not in CI — the agent fails unpredictably. Running Claude's commands inside Docker gives it a consistent environment and makes its behavior reproducible.

Tests as guardrails, not afterthoughts. One engineer put it plainly: "We're adding a lot of tests because we didn't have them — to make sure nothing breaks when Claude is making changes." Without test coverage, agentic changes are a gamble. With it, you have a mechanism that catches regressions before a human has to.

GitHub Actions that close the loop. The goal is for Claude to push a change, trigger a CI run, read the result, and self-correct. That only works if your workflow files are clean and Claude can interpret the output. This pushed the team to write better, more explicit workflow definitions than they'd had before.

Integrated issue tracking. One engineer connected JIRA and GitHub directly so Claude could read the ticket, find the relevant code, and make the change as a single workflow. The observation: "It helps a lot — the context is already there. I don't have to re-explain the task."

None of these are AI-specific breakthroughs. They're good engineering practices that become load-bearing when an agent is doing the work.

---

Three Engineers, Three Different Relationships with Claude

What struck us: "AI-augmented" looked meaningfully different depending on the domain.

On the data acquisition side, one engineer built a system where Claude generates fully functional web scrapers from a recipe library — structured descriptions of platform strategies, API signatures, and scraping hints. Given a target site, the agent selects a strategy, prefers direct API access over proxy scraping when available, and outputs pull requests for human review.

The engineer described his role as "agent wrangler." At peak, this work was consuming roughly $500/month in tokens, the highest on the team, because the agent explores many solution paths per scraper. The next phase: a monitoring agent that tests scrapers on a schedule, files GitHub issues for failures, and triggers a repair agent, all with human review at each gate.

On the backend and data pipeline side, another engineer was decomposing a matching monolith into independent services and building every new repository as agentic-friendly from day one. The framing was explicit: don't retrofit these conventions later. Establish them at the start so future Claude work doesn't require structural rework. The scale of this refactor a meaningful architecture change that previously would have required additional engineers — became tractable because the team applied AI assistance consistently across the codebase.

On the full-stack layer, a third engineer committed to Claude Code as his primary development environment, deliberately minimizing time spent outside of it. He was applying the agentic checklist in real-time, simultaneously onboarding a new team member who was introduced to Claude Code as the default not a later add-on. The new hire learned the conventions as baseline, not exception.

---

The Thing That Didn't Change

Across all three roles, one thing remained constant: the human is always in the loop.

Claude generates artifacts: pull requests, scrapers, workflow definitions. A person reviews and approves them. The engineers weren't describing a system where Claude ships to production unsupervised. They were describing a system where Claude does the drafting. The engineers do the judgment.

This distinction matters more than it might seem. A lot of anxiety about agentic development conflates "Claude handles more of the mechanical work" with "Claude replaces the engineer." Those are fundamentally different things. What we observed was something different: the engineers' time shifted toward higher-impact work: reviewing, guiding, catching errors, setting direction; while Claude handled the mechanical generation.

The analogy that stuck with us: the shift from writing every line yourself to being an editor who maintains quality across a larger surface area.

---

What We'd Tell Your Team (and our other teams)

If you want to move from Claude-as-autocomplete to Claude-as-collaborator, the investment is structural. You're not waiting for better AI, the capability is already there. What you're doing is making your codebase legible to an agent.

Start with the file size rule. It's the highest-impact change with the most visible payoff. Break up your largest files and watch Claude's suggestions get sharper. Then add Docker isolation for any commands you want Claude to run. Then look at your test coverage.

The teams that will get the most out of agentic development over the next few years aren't the ones who wait for a turnkey solution. They're the ones building the habits and conventions now so they can ship products they're proud of, while the rest of their industry is still treating Claude like a smarter tab-complete.

---

Admios is a nearshore engineering partner. We hand-select engineers who seamlessly integrate with US-based technology teams — and yes, we geek out on AI-augmented development practices across every engagement.