<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Ai-Orchestration on Corey Daley</title><link>https://coreydaley.dev/tags/ai-orchestration/</link><description>Recent content in Ai-Orchestration on Corey Daley</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 23 Mar 2026 19:35:00 -0400</lastBuildDate><atom:link href="https://coreydaley.dev/tags/ai-orchestration/rss.xml" rel="self" type="application/rss+xml"/><item><title>Polyphon's MCP Server Makes Your AI Ensemble a Native Agent Tool</title><link>https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/</link><pubDate>Mon, 23 Mar 2026 19:35:00 -0400</pubDate><guid>https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/</guid><description>&lt;p&gt;Polyphon just shipped MCP server support, and it reframes what the tool actually is. You&amp;rsquo;ve been opening it as a UI. Now your coding agent — Claude Code, Cursor, Codex CLI, GitHub Copilot — can call into it directly and broadcast a prompt across your entire Claude + GPT-4o + Gemini ensemble in a single tool call.&lt;/p&gt;
&lt;p&gt;The practical win is not just convenience. It is that you can package disagreement: three independent models, each with different priors, applied to the same question at once. Five tools, two CLI flags, one Settings toggle. The SessionEventSink abstraction is why this works without duplicating orchestration logic — desktop UI and headless agent calls run on the same engine.&lt;/p&gt;
&lt;p&gt;Where in your workflow would paying for disagreement actually change a decision, rather than just produce a louder answer?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/&lt;/a&gt;
&lt;/p&gt;</description></item><item><title>I Built a Tool So AI Models Could Talk to Each Other</title><link>https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/</link><pubDate>Mon, 16 Mar 2026 12:30:00 -0400</pubDate><guid>https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/</guid><description>&lt;p&gt;Every AI power user I know runs the same manual workaround: ask Claude, ask GPT, copy the interesting parts of each into the other, then try to synthesize what you learned. The models are good. The coordination is not.&lt;/p&gt;
&lt;p&gt;I just shipped Polyphon v0.1.0-alpha.2 — a free, local-first desktop app that puts multiple AI voices in the same conversation so they can actually respond to each other. You&amp;rsquo;re the conductor. They&amp;rsquo;re the ensemble. Save a group of voices as a composition and reuse it whenever you need that ensemble again.&lt;/p&gt;
&lt;p&gt;What should a multi-agent conversation feel like when you&amp;rsquo;re not building a pipeline — when you just want to think out loud with several models at once?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/&lt;/a&gt;
&lt;/p&gt;</description></item></channel></rss>