Agent Shortlist

Article

AI Agent Skills and Memory: How to Make Agents Get Better Over Time

Skills files, context management, and routines turn a one-trick agent into a system that improves. Why dumping your whole wiki in on day one is a mistake.

By Lucas Powell·April 29, 2026·6 min read·1,410 words

Your agent is amnesiac.

Every time you start a new session, it wakes up with no memory of you. The agent that helped you craft a sharp client proposal on Tuesday has no idea who you are on Wednesday. Same tool, blank slate. It doesn't remember your tone, your client's name, or the format you spent twenty minutes correcting last week.

This isn't a bug. It's how context windows work. But it means that if you want an agent that genuinely gets better over time — one that knows your preferences, follows your rules, handles recurring tasks without hand-holding — you have to build that deliberately. Nobody does it for you.

Here's how.


The context bloat mistake

The first instinct, when setting up an agent, is to give it everything.

Upload the 200-page SOP. Paste in the last year of customer emails. Attach the full CRM export. You want it to know things, so you feed it everything you've got.

This is the wrong move.

Context windows aren't free, and an agent drowning in information produces worse output than one given the right information at the right time. Irrelevant context crowds out the useful kind. The model can't prioritize what matters if everything is marked as important by virtue of being included.

The analogy: it's like handing a new intern your entire Google Drive on day one and saying "read all of this and then help me with emails." They'll spend the week reading. They still won't know what you actually need. And the emails will be generic.

Good context is specific. It's the briefing doc a good freelancer would ask for before starting work — not the filing cabinet behind your desk.


What a skills file is

A skills file goes by a few names depending on your tool: agents.md, soul file, CLAUDE.md, system prompt. The format doesn't matter much. What matters is what goes in it.

Think of it as the standing brief for your agent. Not everything it could ever know — just what it needs to do this job well.

What belongs in a skills file:

  • Your communication style and tone (with examples, not just adjectives)
  • Output formats you want it to follow — length, structure, what to include and exclude
  • Hard rules: things it should always do or never do
  • Context about your business or audience that changes how it should respond
  • Common errors from previous sessions, added as explicit "don't do this" instructions

What doesn't belong:

  • Raw data dumps (CRM exports, full email threads, spreadsheets)
  • Complete documents it might need to reference occasionally — link or summarize instead
  • Anything you'd only need once

The test: if a competent new hire would need to know it on day one, it's in the skills file. If they'd look it up in the wiki when they needed it, it's not.

For a deeper look at how Claude specifically handles skills, see what Claude skills actually are and how they work.


Converting a workflow into a reusable skill

Once an agent is doing a task well, you don't want to re-explain it every time. You capture it.

A skill is the documented version of a task the agent can already do: the prompt structure, the inputs it takes, the steps it follows, the output format, the edge cases it should handle. Once it exists, you invoke the skill instead of rebuilding the workflow from scratch.

Concrete example: weekly competitor pricing report

Without a skill, you spend 10 minutes writing a prompt from scratch, reminding the agent what format you want, which competitors to cover, and what to do when pricing isn't publicly listed.

With a skill, the brief looks like this:

Skill: Competitor Pricing Report
Inputs: [list of competitor URLs], [week number]
Steps:
  1. Check each URL for visible pricing tiers
  2. Note any changes since last report (flag with "CHANGED")
  3. If pricing is hidden, note "quote-based" and flag for outreach
Output: Table — Competitor | Tier | Price | Notes | Changed?
Format: markdown table, max 1 line per tier

You run it in 30 seconds. The output is consistent every week. When something breaks (a competitor redesigns their site, hides their pricing), you update the skill once and the fix carries forward.

This is the compounding effect people talk about with agents — but it doesn't happen automatically. You have to capture the workflow before it can compound. See AI agent workflow patterns for more on building these structures.


Routines and cron jobs

Skills become more powerful when they run on a schedule.

Monday at 8am, the lead research skill runs. Friday at 4pm, the week summary skill runs. The competitor pricing report goes out every Tuesday morning before the team meeting. You didn't do any of it.

This is the shift from "tool I use" to "system that works for me." And it's more accessible than it sounds.

Three paths to scheduled routines:

Lindy (no-code) — Lindy has native scheduling built into its agent builder. Set a trigger time, attach a skill, done. No code required. Best for non-technical operators who want reliable weekly or daily tasks. See the Lindy platform page.

n8n (workflow builder) — Build a workflow with a Schedule trigger node, then route it to your AI agent node. More setup upfront, but you get full control over inputs, outputs, and what happens downstream with the results. Good for anything that needs to touch other tools (Slack notifications, CRM updates, spreadsheet writes). See the n8n platform page.

Claude API / Claude Code (developer path) — Schedule a script via cron or a service like GitHub Actions. Pass the skill as a system prompt, the dynamic inputs as the user message, and pipe the output wherever you need it. Full control, zero platform dependency. Claude Code is the right starting point if you're already comfortable with a terminal.

The no-code route gets you 80% of the value with 20% of the effort. Start there.


The slowly-onboard-it principle

Here's the trap most people fall into: they try to build a perfect skills file on day one.

They spend three hours writing a comprehensive system prompt, covering every scenario, documenting every preference. Then they start using the agent, and half of it is wrong or missing. Back to the drawing board.

The better approach: start minimal and add incrementally.

Run the agent on real work. When it gets something wrong, add that correction to the skills file. When it misses a preference you care about, document it. When you find yourself editing the same type of output for the third time, that edit is a skills file entry waiting to be written.

After a month of real use, the skills file is genuinely useful — because every line in it came from actual friction. The agent hasn't learned anything (it still starts fresh each session). But you've documented what good looks like, and you load that documentation at the start of every session.

It's not machine learning. It's institutional knowledge management. The distinction matters because it changes how you think about the work: you're not training the agent, you're building a brief that gets better every time you use it.

Just getting started with agents? Read this first before going deeper on skills and memory.


How to tell it's working

You don't need a dashboard. You need three signals:

Fewer edits on first drafts. If you're correcting the same types of errors week after week, the skills file isn't doing its job. If you're catching new errors and almost never seeing old ones, it's working.

Shorter prompts. When you've fully transferred context into the skills file, you should be able to give the agent a one-line instruction and get a complete output. If you're still writing long setup paragraphs to explain the basics, those paragraphs belong in the skills file.

Tasks running without you. The clearest signal: a scheduled routine produces output you use without touching it first. That's not luck. That's a skill file and a routine working together.


The agents most people run are amnesiac by default. The agents that actually compound — the ones that save meaningful hours every week — have a skills file that grows with each session, workflows that are documented before they're automated, and routines that run whether you remember to ask or not.

None of it is complicated. It just requires doing the work of documentation that most people skip.

Start with one task you're doing repeatedly. Write down exactly how the agent should handle it. Load it at the start of the next session. Then add one correction.

That's it. Do that for a month.

About the author

Lucas Powell

Lucas Powell

Founder, Growth 8020

Founder of Growth 8020. Started Agent Shortlist as the publication he wished existed when his team had to pick AI tools.