Part 4: Agents and computer use

The last tool group shows the pattern behind several of the previous tools. Agents are chat applications with tools. They can take actions on your behalf. For coding assistants, those actions can include listing files, reading files, writing files, searching a codebase, and running shell commands.

Product names change faster than the pattern. Once you recognize the agent loop, you can compare tools by the actions they can take, the context they can see, and the guardrails they give you.

Agents in coding tools

The public README gives the practical version:

  • Coding assistants and project bootstrappers both use agents for coding
  • Agents have tools such as reading files, writing files, and indexing the codebase
  • The model chooses when to use those tools

In the Snake examples, the difference is visible. A chat app can return a component. A coding agent can read the existing src/ folder, edit App.jsx, create components/SnakeGame.jsx, and run a build command.

That does not remove your responsibility. It changes where you spend time. You spend less time copying code and more time checking diffs, reading the tool output, and deciding whether the direction still matches the project.

In the earlier demos, the same pattern shows up several times:

  • Claude Code lists files, runs npm commands, and edits Vite files
  • Copilot reads the project and refactors App.jsx
  • Cursor looks at the score state, grid rendering, and border behavior
  • Lovable edits a hosted template and publishes a preview

To learn more about agents

The public README points to two learning resources:

Use those if you want to understand function calling and tool use directly instead of only through a coding product.

Anthropic computer use demo

The workshop also points to Anthropic's computer use demo. This demo gives the model a browser-like computer environment instead of only project files.

Run it with Docker and an Anthropic API key:

docker run \
    -e ANTHROPIC_API_KEY=$ANTHROPIC_API_KEY \
    -v $HOME/.anthropic:/home/computeruse/.anthropic \
    -p 5900:5900 \
    -p 8501:8501 \
    -p 6080:6080 \
    -p 8080:8080 \
    -it ghcr.io/anthropics/anthropic-quickstarts:computer-use-demo-latest

The ports expose the demo interfaces from the container. The exact UI can change over time, so check the upstream quickstart if the container output prints a different local address.

After the container starts, Docker runs a Linux desktop environment. Inside that desktop there is an agent that can interact with the UI. The demo page shows a chat interface and a desktop view.

Start by taking screen control and opening Firefox manually, so the agent does not have to handle browser setup. Then release control and ask:

Find me a picture of a cute kitten.

The agent takes a screenshot, decides it needs a search, types a search query, and opens image results. It is slower than a human, but it shows the important capability: the model can operate a GUI when the tool exposes screenshots and input actions.

PR Agent

The public README also lists PR Agent as another agent example. That tool works on pull request review and description rather than app generation.

The same idea applies outside coding. If the tools are spreadsheet actions, the agent can work on a spreadsheet. If the tools are GitHub actions, the agent can work on pull requests and issues. The tool set defines what the agent can do.

This closes the comparison loop. Chat apps generate code from conversation, IDE assistants edit a local project, bootstrappers create hosted prototypes, and agents describe the tool-using pattern behind all of them.

Questions & Answers (0)

Sign in to ask questions