<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Ai-Tools on Corey Daley</title><link>https://coreydaley.dev/tags/ai-tools/</link><description>Recent content in Ai-Tools on Corey Daley</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 23 Mar 2026 19:55:00 -0400</lastBuildDate><atom:link href="https://coreydaley.dev/tags/ai-tools/rss.xml" rel="self" type="application/rss+xml"/><item><title>Polyphon at v0.8.0: The End of the Prototype Phase</title><link>https://coreydaley.dev/posts/2026/03/polyphon-from-alpha-to-v0-8/</link><pubDate>Mon, 23 Mar 2026 19:55:00 -0400</pubDate><guid>https://coreydaley.dev/posts/2026/03/polyphon-from-alpha-to-v0-8/</guid><description>&lt;p&gt;When I shipped Polyphon v0.1.0-alpha.2, the pitch was simple: put multiple AI voices in one conversation and let them respond to each other. That was useful. But early usefulness and long-term trust are not the same thing.&lt;/p&gt;
&lt;p&gt;v0.8.0 is the release where Polyphon crosses that line. The features that made the difference weren&amp;rsquo;t the ones I planned at launch. Voices can now interact with actual files, with per-voice sandboxing and explicit permission categories. Conversation history is encrypted with SQLCipher whole-database AES-256, with optional password protection. FTS5 search turns the archive into working memory you can actually retrieve from. These aren&amp;rsquo;t incremental improvements — they&amp;rsquo;re the features that decide whether a tool stays an interesting experiment or earns a place near real projects.&lt;/p&gt;
&lt;p&gt;What actually makes you trust an AI tool with real work: capability, privacy, or memory?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/03/polyphon-from-alpha-to-v0-8/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/03/polyphon-from-alpha-to-v0-8/&lt;/a&gt;
&lt;/p&gt;</description></item><item><title>Polyphon's MCP Server Makes Your AI Ensemble a Native Agent Tool</title><link>https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/</link><pubDate>Mon, 23 Mar 2026 19:35:00 -0400</pubDate><guid>https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/</guid><description>&lt;p&gt;Polyphon just shipped MCP server support, and it reframes what the tool actually is. You&amp;rsquo;ve been opening it as a UI. Now your coding agent — Claude Code, Cursor, Codex CLI, GitHub Copilot — can call into it directly and broadcast a prompt across your entire Claude + GPT-4o + Gemini ensemble in a single tool call.&lt;/p&gt;
&lt;p&gt;The practical win is not just convenience. It is that you can package disagreement: three independent models, each with different priors, applied to the same question at once. Five tools, two CLI flags, one Settings toggle. The SessionEventSink abstraction is why this works without duplicating orchestration logic — desktop UI and headless agent calls run on the same engine.&lt;/p&gt;
&lt;p&gt;Where in your workflow would paying for disagreement actually change a decision, rather than just produce a louder answer?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/03/polyphon-mcp-server/&lt;/a&gt;
&lt;/p&gt;</description></item><item><title>I Built a Tool So AI Models Could Talk to Each Other</title><link>https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/</link><pubDate>Mon, 16 Mar 2026 12:30:00 -0400</pubDate><guid>https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/</guid><description>&lt;p&gt;Every AI power user I know runs the same manual workaround: ask Claude, ask GPT, copy the interesting parts of each into the other, then try to synthesize what you learned. The models are good. The coordination is not.&lt;/p&gt;
&lt;p&gt;I just shipped Polyphon v0.1.0-alpha.2 — a free, local-first desktop app that puts multiple AI voices in the same conversation so they can actually respond to each other. You&amp;rsquo;re the conductor. They&amp;rsquo;re the ensemble. Save a group of voices as a composition and reuse it whenever you need that ensemble again.&lt;/p&gt;
&lt;p&gt;What should a multi-agent conversation feel like when you&amp;rsquo;re not building a pipeline — when you just want to think out loud with several models at once?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/03/launching-polyphon-orchestrating-multiple-ai-voices/&lt;/a&gt;
&lt;/p&gt;</description></item><item><title>Making Your AI Subscriptions Pay for Themselves</title><link>https://coreydaley.dev/posts/2026/03/making-ai-subscriptions-pay-for-themselves/</link><pubDate>Sat, 07 Mar 2026 14:00:00 -0500</pubDate><guid>https://coreydaley.dev/posts/2026/03/making-ai-subscriptions-pay-for-themselves/</guid><description>&lt;p&gt;A coworker and I were debriefing after an AI Bootcamp when I said the quiet part out loud: &amp;lsquo;I need my AI subscriptions to pay for themselves.&amp;rsquo; Add up Claude Pro, ChatGPT Plus, GitHub Copilot, Cursor, and a research tool, and you&amp;rsquo;re looking at $100+ a month just to stay current.&lt;/p&gt;
&lt;p&gt;The mental shift that changes everything: stop running your AI stack like subscriptions and start running it like equipment. Every tool needs a job. Assign each one to a revenue output, pick one small experiment, and ship something real in two weeks.&lt;/p&gt;
&lt;p&gt;You don&amp;rsquo;t need a hit app — you need $112/month and a closed loop. Are you running your AI tools in consumer mode or operator mode?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/03/making-ai-subscriptions-pay-for-themselves/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/03/making-ai-subscriptions-pay-for-themselves/&lt;/a&gt;
&lt;/p&gt;</description></item><item><title>The AI Divide: When Innovation Amplifies Inequality</title><link>https://coreydaley.dev/posts/2026/02/ai-ethics-resource-gap/</link><pubDate>Fri, 13 Feb 2026 15:45:00 -0500</pubDate><guid>https://coreydaley.dev/posts/2026/02/ai-ethics-resource-gap/</guid><description>&lt;p&gt;We&amp;rsquo;re witnessing something unprecedented: AI tools that can generate content, build applications, and automate creative work at scales previously unimaginable. But there&amp;rsquo;s a catch—the most powerful capabilities often come with price tags that not everyone can afford.&lt;/p&gt;
&lt;p&gt;From individual creators competing for attention to startups facing AI-augmented giants, the ability to pay for advanced AI is becoming a new axis of inequality.&lt;/p&gt;
&lt;p&gt;Is this just another chapter in technological progress, or are we creating a permanent divide between the AI haves and have-nots? What do you think should be done to ensure AI benefits everyone, not just those who can afford it?&lt;/p&gt;
&lt;p&gt;Read more at &lt;a
 href="https://coreydaley.dev/posts/2026/02/ai-ethics-resource-gap/" target="_blank" rel="noopener noreferrer"&gt;https://coreydaley.dev/posts/2026/02/ai-ethics-resource-gap/&lt;/a&gt;
&lt;/p&gt;</description></item></channel></rss>