๐ Click image to zoom in
Product managers start every day the same way โ opening 8-10 tools to figure out what happened overnight. Email, calendar, Jira, Slack, competitive blogs, analytics dashboards. The data exists across all of them, but the triage is manual, repetitive, and fragmented.
What if AI agents could do the overnight triage autonomously โ scanning every source, filtering signal from noise, cross-correlating across systems, and delivering a single prioritized briefing by the time the PM opens their laptop?
Not a dashboard that displays data. An intelligence layer that thinks about it.
This project demonstrates how multi-agent orchestration enables autonomous information gathering and synthesis โ the same pattern that would power internal decision-support tools at any product-led company.
A hierarchical supervisor pattern: one Synthesizer agent (Claude Sonnet) coordinates four specialized worker agents (GPT-4.1 nano), each connected to a different data source. The workers handle high-token ingestion on cheap models; the Synthesizer handles cross-source reasoning on a strong model.
Scans inbox, classifies messages by urgency, filters noise (newsletters, FYIs, cc'd threads), surfaces action items.
Detects schedule changes, cancelled meetings, and time conflicts. Recommends keep/skip/prep for each event.
Monitors ticket queue for escalations, P1 bugs, and status changes. Flags items requiring PM attention.
Runs targeted searches for competitor product updates, pricing changes, and announcements.
Receives condensed outputs from all four workers. Cross-references signals, ranks by priority, identifies correlations across sources, and produces a structured briefing with recommended actions.
The workflow runs on n8n, self-hosted via Docker on a Mac Studio. Sequential execution โ each agent runs in order, results flow up to the Synthesizer. No external orchestration framework, no cloud deployment. The entire system runs locally.
The workflow triggers and each agent independently connects to its data source โ Gmail, Google Calendar, Jira, and Tavily. They ingest raw data using GPT-4.1 nano, classify and filter it, then pass condensed summaries to the Synthesizer.
Claude Sonnet receives all four agent reports and identifies connections between them โ a Jira P1 ticket that matches an escalation email, a calendar meeting that needs prep based on competitive intel. Signals from one source inform the priority of another.
The Synthesizer produces structured JSON โ ranked priority items with recommended actions, calendar events with keep/skip/prep tags, an overnight changelog, and summary statistics. "3 need action" not "47 emails."
The briefing populates a dashboard where the PM can scan everything in under 2 minutes โ compared to 60+ minutes manually triaging across 8-10 tools. Each item links back to its source for drill-down.
To demonstrate the impact of the multi-agent system, we designed a custom PM dashboard in Figma that visualizes exactly what the agents produce. The PM checked in Wednesday at 6:00 PM โ nothing new. Overnight, four agents worked autonomously. By Thursday 8:15 AM, the dashboard shows what changed and what needs attention.
๐ Click image to zoom in
Each annotated screenshot isolates one agent's contribution โ showing how a single detection ripples across the entire dashboard. Numbered markers connect to the summary box.
๐ Click any image to zoom in
Beyond the individual agent reports, the Synthesizer surfaced patterns that would take a PM significant time to piece together manually.
The APAC checkout error appeared in both email (from James Liu at 2:14 AM) and Jira (P1 ticket). The Synthesizer confirmed it was the same incident from two independent sources and escalated it to the top of the priority list. No single tool tells that story alone.
The Email Agent scanned 47 messages and surfaced 4. 43 were noise โ newsletters, FYIs, cc'd threads. The intelligence isn't in showing everything; it's in knowing what to hide. "3 need action" is more useful than "47 unread."
The system didn't just detect that VP Strategy Sync moved to 3:00 PM โ it connected that meeting to the Competitive Intel findings and recommended the PM prep with the competitor pricing data. The calendar change informed the priority list, which informed the meeting prep. One change, traced across four sections.
This demo uses seeded test data across Gmail, Google Calendar, and Jira. In production, the agents would connect to the same APIs with real data. The architecture is identical โ only the volume and variability of data changes.
Agents run one after another for easier debugging and deterministic behavior. For a daily batch workflow, sequential is sufficient. Parallel execution would reduce total runtime but adds coordination complexity โ a worthwhile optimization for real-time use cases.
The frontend is a design artifact, not a working application. The value proposition is the intelligence layer โ the multi-agent orchestration, cross-source reasoning, and structured output. The dashboard is a commodity; building it in React would demonstrate frontend skills, not AI product thinking.
The Email Agent may over-prioritize messages with urgent keywords without sufficient context. The Competitive Intel Agent may flag irrelevant posts due to keyword similarity. The Synthesizer may overweight certain sources depending on prompt framing. Mitigations include structured output schemas, confidence scoring, and threshold filtering.
The current system demonstrates the core pattern with 4 agents. In production, this architecture extends to any data source a PM touches.