2/12Back to guide
Part 1: Foundations

Coding Agents vs Code Editors

Shawn Tenam·12 min read·2,489 words

I was five months into my SDR job when I hit the wall.

Not the "this is hard" wall. The "I am manually doing the same twelve steps for every single prospect and there has to be a better way" wall. I had a spreadsheet with 200 target accounts. For each one, I was opening LinkedIn, finding the right VP, checking their company's recent news, cross-referencing against our CRM to make sure nobody else had touched them, writing a personalized first line, and logging the whole thing. Each account took 15-20 minutes of clicking between tabs.

I asked my manager if there was a tool for this. He pointed me to our sequencing platform. It helped with the sending part, sure. But the research, the personalization, the decision of who to reach out to and why right now? That was still on me. Still manual. Still slow.

That was before I understood the difference between tools that help you type and tools that help you build. It is the most important distinction in this entire playbook, and most GTM operators get it wrong because the marketing around AI tools blurs the lines on purpose.

So let's draw them clearly.

The Three Categories

Every AI tool you will encounter as a GTM operator falls into one of three buckets. Understanding which bucket you are reaching into determines whether you get results or waste an afternoon wondering why AI feels overhyped.

Chat interfaces

ChatGPT, Claude.ai, Gemini. You type a question, you get an answer. They are good at answering quick GTM questions ("What's a typical reply rate for cold outbound to VPs of Engineering?"), brainstorming subject lines, explaining concepts you have not encountered before, and summarizing documents you paste in.

What they cannot do: read your files, run scripts, modify a codebase, call APIs, or remember what you told them last Tuesday. Every conversation starts from zero unless you manually paste context back in. They are a very smart coworker with amnesia.

Here is what this looks like in a GTM workday. You are about to hop on a discovery call with a prospect you have not researched yet. You paste their company's About page into Claude.ai and ask, "What are three questions I should ask based on this?" You get useful answers in 30 seconds. That is the right use case. Quick question, quick answer, move on.

Where it falls apart: you try to use the same chat window to build an entire outbound workflow. You paste your ICP in one message, your positioning in the next, your competitor list in the next. By message twelve, the AI has lost track of what you told it in message three. You are fighting the tool instead of using it.

Use chat interfaces for: Quick questions, brainstorming, one-off copy tasks. Any code comfort level.

AI code editors

Cursor, Windsurf, GitHub Copilot. These are code editors with AI bolted in. You open a project folder, the AI can see your files, and it helps you write and edit code through a visual interface. They are good at editing files you can see in the sidebar, autocompleting code as you type, explaining code when you highlight it, and making changes across a few files when you point them in the right direction.

The key advantage is the GUI. You see your file tree, your open tabs, your terminal output. If you have never touched code before, this visual scaffolding matters. You can watch the AI suggest a change, read the diff, and approve it before anything lands. It is like pair programming with a senior engineer who does not judge you for not knowing what a function is.

For GTM, this is where a lot of people start and where many stay longer than they should. You open Cursor, create a project folder for your GTM workspace, and start building files. The AI helps you structure your ICP document. It autocompletes your competitor analysis. It suggests edits to your email templates. It is genuinely useful.

But you are still driving. You are still the one opening files, telling the AI which ones to look at, approving each change one at a time. For editing an existing file or learning your way around a codebase, this is exactly right. For building a multi-step automation that pulls data from three APIs and generates personalized outreach? You are doing the coordination work that the tool should be doing.

Use AI code editors for: Learning your way around a codebase, making targeted edits, anything where you want to see changes before they happen. Best for code comfort levels 1-2.

Coding agents

Claude Code, Codex CLI, Aider. These are autonomous agents that operate in your terminal. You give them a task. They read whatever files they need, run whatever commands are required, create or modify files, and execute multi-step plans without you pointing at each file individually.

Here is what that looks like in practice:

# You open Claude Code in your project folder
claude

# You type a task
> Read my ICP file, pull the top 50 companies from Apollo that match,
> enrich them with Exa for recent news, and save the results as a CSV.

The agent then reads gtm-os/demand/icp.md to understand your target buyer, writes a Python script that calls the Apollo API, runs the script and gets the results, writes another script to enrich those results with Exa, and saves the final output as a CSV.

You did not open a single file manually. You did not write a line of code. You described the outcome, and the agent figured out the steps.

This is the difference between driving and having a driver. The AI code editor is power steering. The coding agent is an autonomous vehicle. Both get you from A to B. One requires your hands on the wheel the whole time.

Use coding agents for: Multi-step GTM workflows, building automation, anything where the task requires reading context, making decisions, and executing. Best for code comfort levels 3-4, though 2s can learn fast with structured context.

The Real Tradeoff: Power vs Accessibility

The tradeoff between coding agents and AI code editors is not "better vs worse." It is power vs accessibility.

AI Code Editor (Cursor) Coding Agent (Claude Code)
Interface Visual GUI with file tree, tabs Terminal, text in, text out
Autonomy Suggests edits, you approve each one Executes multi-step plans
Context Sees open files, sometimes the whole project Reads any file it needs, when it needs it
Learning curve Low, feels like VS Code Medium, requires terminal comfort
Best for Editing existing files, learning Building new things, automation

If you have never used a terminal, start with Cursor. Read these chapters. Practice. When you find yourself thinking "I wish I could just tell it to do the whole thing," that is when you are ready for Claude Code.

If you are already comfortable in a terminal, skip straight to Claude Code. You will move faster.

Real GTM Scenarios Across All Three

Let me make this concrete with five scenarios that come up every week in GTM work.

Scenario 1: Pre-call research. You have a discovery call in 10 minutes. You need to know what the prospect's company does, their recent funding, and two good opening questions. This is a chat interface job. Paste their website URL or About page into Claude.ai, ask your questions, get answers, hop on the call. Time: 2 minutes.

Scenario 2: Updating your ICP document. You just closed three deals that do not match your existing ICP. You need to revise your target criteria based on what is actually working. Open Cursor, navigate to your ICP file, tell the AI "update this ICP based on these three recent wins" and paste the deal summaries. Review the suggested changes, approve them. This is an AI code editor job. You want to see and control each edit. Time: 15 minutes.

Scenario 3: Building a new outbound sequence. You have a new product launch and need a 5-email sequence targeting CTOs at enterprise companies. A coding agent reads your ICP, your positioning doc, your voice profile, and your competitors file, then generates all five emails with personalization hooks, subject lines, and follow-up logic. You review the output. This is a coding agent job. Time: 10 minutes.

Scenario 4: Weekly pipeline enrichment. Every Monday, you need to pull new companies from Apollo that match your ICP, enrich them with recent news, score them, and push the top 20 into HubSpot as new contacts. This is absolutely a coding agent job. You build the workflow once, and it runs every Monday. With a chat interface, you would spend an hour manually doing this every week. With a coding agent, you spend 30 minutes building it and then it is automated.

Scenario 5: Analyzing competitor messaging. You want to understand how three competitors position against each other and find the gap your product fills. A chat interface can do this if you paste in the competitor pages. An AI code editor can do it if you have the data in files. A coding agent can do the whole thing: scrape the competitor pages, analyze the positioning, compare against yours, and write a gap analysis document. Same task, three levels of leverage.

The pattern: as you move from chat to editor to agent, each step multiplies your output while reducing your manual coordination. The chat interface requires you to do all the fetching and pasting. The editor requires you to point at files. The agent figures out what it needs on its own.

What Changes When You Think Like an Operator

When I was an SDR, I thought AI was about writing better emails. I would paste a prospect's LinkedIn headline into ChatGPT and ask for a personalized first line. It was fine. Marginal improvement over what I was already doing. Maybe it saved me 3 minutes per prospect.

The real shift happened when I stopped thinking about AI as a writing tool and started thinking about it as an execution layer. I did not need a tool to write a slightly better cold email. I needed a tool that could take my ICP, find the right companies, check their recent news for relevance signals, draft personalized outreach based on those signals, and queue the whole thing up for me to review. That is not a prompt. That is a workflow. And that is what coding agents do.

This is the mental model shift. You are not "prompting." You are giving an agent a workspace and instructions.

When you use ChatGPT, you are prompting. You craft a message, you get a response, you craft another message. It is a conversation.

When you use a coding agent, you are operating. You build a workspace (files, folders, CLAUDE.md, scripts), and the agent operates inside it. The quality of the workspace determines the quality of the output, not the cleverness of your prompt.

This is why GTM operators who dump everything into a single chat message get mediocre results. They are treating an agent like a chatbot. The agent does not need a better prompt. It needs a better workspace.

Mediocre: "Write me a cold email for VP of Engineering at mid-market SaaS companies"

Better:   A workspace with an ICP file, positioning doc, competitor analysis,
          voice profile, and a CLAUDE.md that says "when asked to write
          outbound emails, reference all demand/ files and use voice/profile.md"

The first approach gives you generic output that sounds like every other AI-generated email. The second gives you output that sounds like you, targets the right buyer, addresses the right pain points, and avoids your competitors' talking points. Same AI. Different workspace.

CLAUDE.md: The Universal Entry Point

Here is the thing that ties this all together: CLAUDE.md.

When you open Claude Code (or Cursor, or any agent that supports it) inside a project folder, the first thing it does is read a file called CLAUDE.md in the root directory. This file tells the agent who it is and what it should do, what the project structure looks like, rules it must follow, and domain knowledge it needs.

Think of it as the agent's job description. You would not hire someone and then explain the role from scratch every time they walk into the office. You give them an onboarding doc. CLAUDE.md is that onboarding doc.

We will go deep on this in Chapter 02. For now, just know that it exists and that it is the single highest-leverage file in any GTM workspace. A good CLAUDE.md turns a general-purpose AI into a GTM operator that knows your business, your buyers, and your voice. A missing CLAUDE.md means the agent is starting cold every single time.

A Typical GTM Workday With All Three Tools

Here is how a GTM operator who understands these categories structures their day:

Morning (batch processing with coding agents). Start the day by running your automated workflows. "Pull this week's intent signals and match them against my target account list." "Enrich the 15 new inbound leads from yesterday." "Generate personalized first lines for my Monday send list." These are all coding agent tasks. You describe the outcome, the agent executes, you review the results over coffee.

Midday (precision editing with AI code editors). After reviewing the agent's output, you make targeted edits. This email's tone is off. This account's pain point is wrong. This sequence needs a different CTA for the enterprise segment. Open Cursor, make surgical changes, approve each one. Editors are perfect for this kind of refinement work.

Throughout the day (quick answers with chat). A prospect asks about a feature you are not sure about. You paste the question into Claude.ai with some product context and get a draft response in 30 seconds. Your manager asks for a quick competitive comparison. You ask the chat interface. A candidate for your team asks what "GTM engineer" means. Quick question, quick answer.

End of day (setup and planning with coding agents). Before you log off, you set up tomorrow's workflows. "Update my competitor analysis with any new messaging changes from this week." "Draft a content outline based on the three deals we lost this month." "Build a new email sequence for the product update launching Thursday." The agent works through these while you write up your notes.

The key insight: each tool has a natural place in the workflow. The mistake is trying to use one tool for everything. Chat interfaces for building automation is painful. Coding agents for a quick one-off question is overkill. Match the tool to the task.

Where to Go From Here

If you are at code comfort level 1-2, your starting tool is Cursor. Read every chapter in this playbook. The concepts (context engineering, token efficiency, structured workspaces) apply to any tool. When you feel limited by the editor, move to Claude Code.

If you are at code comfort level 3-4, install Claude Code. Chapter 02 is the highest-leverage thing you can learn next. It is about context engineering, which is the skill that separates people who get magic from coding agents from people who get mediocre autocomplete.

If you already have a CLAUDE.md and a structured workspace, you are ahead of most. Jump to Chapter 03 to optimize your token usage, then straight to whatever GTM workflow you need.


For the hands-on version with exercises, see Chapter 01 on GitHub.

Next: Chapter 02 - Context Engineering -- How to structure the context that makes your agent actually useful.

Open Source Playbook

This guide is open source. Fork the repo to get the hands-on version with exercises, templates, and a full GTM-OS skeleton.

ShawnOS.ai|theGTMOS.ai|theContentOS.ai
built with Next.js · Tailwind · Claude · Remotion