Case Study

How I Built a 4-Agent System That Runs a Social Media Agency

Camille is a social media manager running 4 to 7 clients at the same time. Every Monday she needed briefings for each one. Every month she needed reports for each one. Every week she needed content ideas, captions, hashtag sets, and posting schedules across multiple platforms. All of this done manually, in parallel, without dropping a client or losing a brand voice.

The workload was not unsustainable. But it was compressing the time she had available to think clearly, build relationships, and actually grow her business. She was spending most of her hours on work that is high-effort but low-judgment: gathering information, formatting reports, generating variations on content that follows a pattern she already knows well.

That is where I came in.


The problem

The challenge with social media management at scale is not that any single task is hard. It is that there are many of them, they repeat on tight schedules, they each require different brand voices, and they need to sound like a person, not a system.

Generic AI tools fail here for a specific reason: they produce output that reads like AI. The moment a client sees a caption that could belong to anyone, trust breaks. The whole point of a social media manager is consistency with personality. An AI system that can not embed that personality is worse than no AI at all, because it creates more editing work than starting from scratch.

The goal was not to replace Camille. It was to free her from the repetitive parts so she could focus on the judgment calls: client strategy, relationship building, creative direction, and growing her agency.


Why build custom instead of using a tool

The market has many social media AI tools. Camille had tried several. The problem was consistent: they were designed for generalist use, not for a multi-client agency workflow where each client has deeply different tone, audience, and content rules.

Off-the-shelf tools could not hold per-client memory. They could not pull from Camille's own notes and templates. They could not be scheduled to fire automatically every Monday morning. And they produced output she had to rewrite anyway.

The decision to build custom came down to this: the time saved by having a system that actually works pays for the build in weeks, not months. If you are evaluating a similar move, the AI Automation Systems service page lays out exactly what that build looks like.


The 4-agent architecture

The system is built with four Claude Code agents, each with a specific role, a defined memory scope, and a clear set of tools. They run inside Camille's Obsidian vault, which already contained her client folders, brand notes, and weekly logs.

Obsidian was the right workspace for this because it is file-based. Every note is a markdown file. Every client folder is a directory. Agents can read and write directly, which means Camille sees the output in the same place she does her thinking, without learning a new interface.

The four agents are:

  1. Weekly Briefing Agent: fires every Monday at 8 AM, reads each client's folder, generates a structured briefing for the week
  2. Content Pipeline Agent: generates platform-specific captions, variants, and hashtag sets from a brief or topic list
  3. Client Report Agent: takes performance notes and compiles structured monthly reports per client
  4. Research and Monitoring Agent: tracks industry news, trend signals, and competitor context relevant to each client's sector

Each agent has read access to the shared vault and write access only to its designated output folder. This keeps the system clean and makes it easy to review what each agent produced before anything goes to a client.


Agent 1: the Weekly Briefing Agent

The briefing agent runs on a cron schedule every Monday at 8 AM. It reads each client's Brand Notes file, the previous week's performance log, and any open tasks tagged #this-week in the client folder.

From those inputs it generates a structured briefing: what performed well last week, what to focus on this week, 3 to 5 content angles to explore, and a reminder of any upcoming events or campaigns for that client.

The key to making the briefing useful is the system prompt. It is not generic. It embeds Camille's own briefing format (the structure she had already developed over years) and instructs the model to write as a strategist who knows this client, not as a content machine. The output reads like something Camille would have written after spending 40 minutes reviewing the account.

Before the system, Camille spent around 20 to 30 minutes per client writing Monday briefings. With 5 active clients, that was 2 hours before 10 AM. Now the briefings are waiting for her when she opens the vault. She reviews, adjusts where needed (usually small things), and moves to the actual work.


Agent 2: the Content Pipeline Agent

The content pipeline agent takes a topic list or content angle and generates full platform-specific posts: Instagram captions with line breaks and emoji placement, LinkedIn text with professional framing, Twitter/X threads if relevant.

The critical part is the per-client voice instruction. Each client has a Voice.md file in their folder: tone descriptors, words they use and words they never use, energy level, formality, reference examples of posts that worked well. The agent reads this before generating anything.

This is what separates the output from generic AI content. The model is not guessing what the brand sounds like. It has been given a precise specification. The result is first drafts that need light editing, not a full rewrite.

Caption variants are generated automatically: 3 per post, each with a different angle (educational, behind-the-scenes, direct offer). Camille picks one or combines elements from two. The hashtag sets are pulled from a Hashtags.md file she maintains per client, refreshed monthly.


Agent 3: the Client Report Agent

Monthly reporting was one of the biggest time drains. Pulling data from platform insights, formatting it, writing commentary, making it look professional: a few hours per client, every month.

The report agent takes two inputs: a raw notes file where Camille logs performance observations during the month, and a template she designed for each client (which varies slightly by client preference). It generates a full report draft in Obsidian markdown, ready to be exported or shared.

The commentary section is what makes this work. Generic performance reports are just numbers. The agent writes actual analysis: what the numbers mean in context, what changed compared to the previous period, what to adjust next month, and a clear summary the client can read in 3 minutes without a social media background.

This only works because the system prompt includes Camille's own commentary style and the client context. The model knows what industry the client is in, what their goals are, and how they prefer to receive feedback. The output sounds like Camille wrote it after sitting down with the data for an hour.


Agent 4: the Research and Monitoring Agent

The research agent runs weekly and produces a short sector briefing for each client: 3 to 5 developments in their industry, 1 or 2 trend signals worth tracking, and any notable competitor activity. It pulls from a curated source list and summarises findings into the client's research folder.

This agent is the most opinionated: it filters out noise, flags only what is genuinely relevant to the client's positioning, and frames each finding as a potential content or strategy angle. It does not just summarise the news. It connects it to the client's context.

The output feeds directly into the Monday briefing. The briefing agent reads the research folder as one of its inputs. This creates a chain: research informs briefing, briefing informs content, content goes to the client.


Keeping the human voice

This was the non-negotiable requirement. Camille's clients hired her, not a system. If the content started sounding like everyone else's AI output, the business would suffer.

The solution is multi-layered. First: the Voice.md specification per client, with real examples and anti-patterns. Second: explicit instructions in every system prompt to avoid AI-sounding constructions (phrases like "dive in," "delve into," "it is worth noting," excessive em dashes, hollow openings). Third: Camille's review step, which takes 5 to 10 minutes per client rather than 20 to 40, and focuses on feel rather than structure.

The goal was never to remove the human from the loop. It was to make the human's time in the loop count for more.


Results after deployment

Three months in, the system is running in production across all of Camille's active clients. The numbers are straightforward:

  • 70 to 80% reduction in time spent on weekly briefings, content generation, and monthly reports
  • Monday briefings waiting in the vault before she opens her laptop
  • Monthly reports cut from 3 to 4 hours of manual build time to under 45 minutes of review and export per client
  • Content output consistent across clients, with brand voice maintained and no client complaints
  • Headspace freed for strategy, client calls, and business development

The less obvious result: Camille took on a new client shortly after deployment. Before the system, that would have required hiring. With the system, it required adjusting a few config files and adding a new vault folder.


What this means for your business

The Camille system is a specific solution to a specific problem. But the pattern is universal: if your business has high-volume repetitive work that requires consistency and brand awareness, the same approach applies.

The key ingredients are:

  • A clear picture of what is being automated and what stays human
  • Per-entity memory (per client, per product, per context) not a single generic prompt
  • A workspace the user already uses (do not add friction by making them learn a new tool)
  • Scheduled execution so the system delivers without anyone having to trigger it
  • A review step that is fast, not optional

If you run a service business and you are spending hours on work that follows a pattern you already know well, you are a good candidate for something like this. The conversation to figure out if that is true takes 30 minutes. You can see the full Camille system on the Camille case study page, or explore the AI Agent Architecture service for what a build like this covers end to end.

Build Yours

Want a system
like this one?

Book a free 30-minute call. We map your situation, identify the highest-impact automation, and figure out if we are a fit.

Book Free 30-min Call