Agent Shortlist

Article

The 5 most common AI agent use cases (and which platform fits each)

What are builders actually using AI agents for in 2026? Five patterns we see repeatedly, with the platform we'd pick for each — no vendor demos.

By Lucas Powell·April 28, 2026·8 min read·1,782 words

If you're trying to figure out what AI agents are actually for, the marketing layer is useless. Every vendor frames the category around their feature set. The honest answer comes from looking at what builders deploy in production — what they keep using six months later, what gets renewed, what generates real value.

We've watched what builders do with the 22 platforms we've reviewed. Five use cases come up repeatedly. They're not all the obvious ones, and the platform you'd pick for each is different. Here's the map.

1. Autonomous software engineering ("vibe coding")

The fastest-growing use case in 2026. AI agents that build software for non-developers, and AI agents that accelerate work for developers who already know what they're doing.

The pattern: a builder describes what they want — an internal tool, a script, a feature, a fix — and an AI agent reads the relevant code, plans the approach, makes the changes across multiple files, runs tests, and creates a pull request. The whole loop runs without human keystrokes.

For developers, this is straightforward productivity. For non-developers, this is a step change. People who couldn't write code two years ago now ship working internal tools by describing what they want in plain English. The phrase that's stuck is "vibe coding" — and it's a real shift in who can build software.

Concrete examples we've seen work:

  • A founder reading a Linear ticket, having an agent pull down the codebase, write the code, run the tests, and create the PR — completely autonomously
  • An ops lead building a dashboard that pulls from a spreadsheet, processes uploaded files, and emails results — all through plain-English instructions to a coding agent
  • An engineer migrating an entire module from one framework to another in an afternoon instead of a sprint

Which platform: Claude Code is the default — included with a Claude Pro subscription, terminal-first, agentic loop. Cursor wins for IDE-first workflows and multi-model flexibility. Aider for cost-conscious teams who want open-source and bring-your-own-API-keys.

2. Deep research and data aggregation

The use case where AI agents flip the unit economics most dramatically. Tasks that used to require weeks of analyst time complete in days.

The pattern: an agent scours a defined research surface — the web, internal documents, structured data sources — and synthesises the findings into a brief, a report, or a structured dataset. The work that's hard for humans (reading hundreds of sources, cross-referencing, finding patterns) is exactly the work LLMs handle well. The work that's hard for LLMs (calling, qualitative interviews, on-the-ground judgement) doesn't fit this pattern.

Concrete examples that produce real ROI:

  • Stock research that identifies bottlenecks in the AI supply chain by scouring earnings calls, press releases, and analyst reports
  • A roofing company analysing satellite imagery alongside recent hail-storm data to find blue-collar sales leads in affluent neighbourhoods
  • Drug-trial failure prediction by reading hundreds of pages of FDA filings and clinical study PDFs

The pattern that drives the highest ROI here is model routing: cheap models like Claude Haiku 4.5 or DeepSeek V4 Flash do the high-volume scraping and parsing; expensive models like Claude Opus 4.7 are reserved for the final synthesis step. This drives the cost of a research-quality lead down to fractions of a cent.

Which platform: Hermes for technical operators who want a server-deployed agent that builds skills over time. OpenClaw for personal research workflows. Manus AI for non-technical builders who want autonomous browser-based research without configuration. For the model layer, the cost calculator shows the routing math.

3. Automated content and social media pipelines

Where solo creators and small marketing teams are getting the biggest leverage. The pattern looks like a pipeline rather than a single agent.

The pattern: research → write → review → publish, with each step handled by a different agent (or a different prompt to the same agent). Researcher agent identifies trending topics or pulls performance data from your past content. Writer agent drafts. Editor agent reviews against brand voice rules. Publisher pushes to the platform.

What separates this from earlier "AI content tools" is the loop. The writer agent reads what your previous posts performed well or failed, and adjusts. The system gets better at your specific voice and audience over time, not just the generic LLM defaults.

Concrete patterns we've seen work:

  • A solo creator running a four-agent newsletter pipeline (researcher, writer, editor, reviewer) that ships weekly without manual intervention
  • A SaaS company analysing past Threads performance to identify winning post formats, then drafting new posts in those formats automatically
  • Marketing teams using frameworks like Remotion to write code that generates animated data visualisations and promotional videos — animations as code, written by an agent

Which platform: Lindy for the no-code path with polished templates. Relevance AI for more sophisticated multi-step workflows with conditional logic. n8n for teams with a developer who want full control over the pipeline. For coding-driven content workflows (like the Remotion pattern), Claude Code sits underneath the agent.

4. Customer support and outbound voice agents

The most rapidly maturing category in 2026 — and the one where AI agents move from "drafting" to "actually doing" with real customers.

The pattern: an AI agent handles voice conversations end-to-end. Inbound support calls (order status, password resets, FAQ deflection) or outbound sales calls (lead qualification, appointment booking, renewal follow-ups). With sub-second latency, customers often don't realise they're talking to an agent until later — and increasingly, they don't care, as long as the conversation actually solves their problem.

Two non-obvious patterns we keep seeing:

  • Dynamic language switching mid-call. YouTube TV's support uses agents that detect a customer's preferred language from the first few seconds and switch the agent's voice and language accordingly. Native speakers, multiple languages, no human agent matching needed.
  • Outbound at SDR-replacement scale. Voice agents running thousands of qualification calls per day, with deflection rates that make the per-minute cost (~$0.07–0.15) negligible against the human SDR equivalent ($6,000+/month).

Which platform: Retell AI for builder-friendly production voice agents. Vapi for developer-first customisation. Bland AI for high-volume outbound campaigns. ElevenLabs Conversational when premium voice quality is part of your brand. The full breakdown is in the Voice AI section of the shortlist.

5. Personal administration and "second brains"

The least-talked-about use case publicly, but the one many builders rely on most heavily for their own work. AI agents as 24/7 personal assistants — not "Siri" assistants, but agents that read, categorise, summarise, and act on the digital noise that fills a builder's day.

The pattern: agents running on your own infrastructure (or in your own Claude account), reading your email, your messages, your calendar, your fitness data, your notes. They categorise, summarise, surface what matters, and act on the routine stuff. A wiki-style notes layer accumulates over time so the agent gets better at understanding your priorities, your projects, your people.

Concrete examples we've seen builders rely on:

  • An agent that reads every email at 7am, files routine ones, surfaces 3–5 that need real attention, and drafts replies for the 3–5
  • A morning briefing agent that scrapes Apple Health data, summarises last night's sleep and yesterday's activity, and includes it alongside the day's calendar and important emails
  • A persistent personal wiki where notes accumulate across years, queryable by an agent that knows your context — "what did I think about that vendor in March?" gets a real answer

Which platform: OpenClaw for the personal AI harness — installs on your machine, connects to messaging platforms (WhatsApp, Telegram, iMessage), runs continuously. Hermes for server-deployed always-on agents that improve over time. Claude Code if your "personal admin" includes building automation scripts you actually use yourself.

The cross-cutting pattern: model routing

The single biggest cost-optimisation pattern across all five use cases is model routing — using cheap models for high-volume mechanical work and expensive models for the few decisions that matter.

A research workflow that runs the full Claude Opus 4.7 model on every web page it scrapes is paying frontier-tier prices for the equivalent of OCR. Routing the scraping to Haiku 4.5 ($1/M input) and reserving Opus ($5/M input) for the final synthesis cuts costs 80%+ with no real quality loss.

The teams getting the highest ROI from AI agents in 2026 are not the ones using the smartest models everywhere. They're the ones who routed work to the cheapest model that handles it, and only escalated when the task actually needed it.

The cost calculator shows the spread for any workflow you choose — usually 50–100× between the cheapest and most expensive model for the same task. That's not a tax to pay; that's a budget to optimise.

Where AI agents still fail

The honest middle. Five use cases work in 2026; many others still don't.

What we'd be cautious about deploying agents for unsupervised:

  • High-stakes client communications — escalations, complaints, contract negotiations
  • Financial decisions without a human review step — anything involving money, anything regulated
  • Real-time emotional intelligence — therapy, conflict resolution, sales calls into a hostile prospect
  • Tasks where a wrong answer is irreversible and discovered late — legal filings, medical guidance

The pattern that works in 2026: AI agents as force multipliers for a human, not autonomous replacements. Handle the first 80% of the task; keep a human on the final call. Going fully autonomous on the remaining 20% is rarely worth the risk.

Where to start

If you're at the "I should use AI agents but I don't know where" point, the order we'd recommend:

  1. Try a coding agent for one workflow you keep retyping. Even non-developers benefit from this — describe a tool you wish existed, watch Claude Code build it. The bar to entry is the lowest of any category.
  2. Add one no-code agent for a sales or support workflow. Lindy for the fastest path to working. The ROI calculation against your existing process makes itself.
  3. Add a research agent if your work involves reading more than writing. Hermes or OpenClaw if you're technical, Manus AI if you're not.
  4. Defer voice agents until you've nailed text-based ones. Voice has higher infrastructure complexity; the text patterns transfer when you're ready.

If none of those feel right, run the five-question Agent Picker. It asks the questions that surface which use case fits your team and recommends a starting platform.

The 22-platform Shortlist has the full opinionated rankings if you want to skip the picker and read at length.

About the author

Lucas Powell

Lucas Powell

Founder, Growth 8020

Founder of Growth 8020. Started Agent Shortlist as the publication he wished existed when his team had to pick AI tools.