<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>LLM Agents on SPERIXLABS</title><link>https://sperixlabs.org/series/llm-agents/</link><description>Recent content in LLM Agents on SPERIXLABS</description><generator>Hugo</generator><language>en-us</language><copyright>SPERIXLABS</copyright><lastBuildDate>Wed, 15 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://sperixlabs.org/series/llm-agents/index.xml" rel="self" type="application/rss+xml"/><item><title>LLM-Redactor: What Leaves Your Prompt When You Talk to a Cloud LLM</title><link>https://sperixlabs.org/post/2026/04/llm-redactor-what-leaves-your-prompt-when-you-talk-to-a-cloud-llm/</link><pubDate>Wed, 15 Apr 2026 00:00:00 +0000</pubDate><guid>https://sperixlabs.org/post/2026/04/llm-redactor-what-leaves-your-prompt-when-you-talk-to-a-cloud-llm/</guid><description>&lt;p&gt;Every time a coding agent sends a prompt to a cloud LLM, the full content of that prompt &amp;mdash; your code, your credentials, your customer names, your internal project codenames &amp;mdash; lands on someone else&amp;rsquo;s server. It may be logged, retained for training, produced in response to subpoena, or exfiltrated in a breach. TLS protects the wire. Nothing protects the content.&lt;/p&gt;
&lt;p&gt;We built &lt;strong&gt;&lt;a href="https://github.com/jayluxferro/llm-redactor"&gt;LLM-Redactor&lt;/a&gt;&lt;/strong&gt; to measure exactly how much leaks and what you can do about it. The &lt;a href="https://arxiv.org/abs/2604.12064"&gt;paper&lt;/a&gt; evaluates eight techniques on a common benchmark. This post is the practitioner&amp;rsquo;s summary.&lt;/p&gt;</description></item><item><title>Local-Splitter: Cutting Cloud LLM Costs by Putting a Small Model in Front</title><link>https://sperixlabs.org/post/2026/04/local-splitter-cutting-cloud-llm-costs-by-putting-a-small-model-in-front/</link><pubDate>Wed, 15 Apr 2026 00:00:00 +0000</pubDate><guid>https://sperixlabs.org/post/2026/04/local-splitter-cutting-cloud-llm-costs-by-putting-a-small-model-in-front/</guid><description>&lt;p&gt;Cloud LLM tokens are expensive. Not in the &amp;ldquo;my AWS bill is high&amp;rdquo; sense &amp;mdash; in the &amp;ldquo;I&amp;rsquo;m burning $0.015 per 1K output tokens and my coding agent sends 200+ requests per session&amp;rdquo; sense. Most of those requests don&amp;rsquo;t need a frontier model. &amp;ldquo;What does this function return?&amp;rdquo; doesn&amp;rsquo;t need Claude Opus. &amp;ldquo;Add a docstring here&amp;rdquo; doesn&amp;rsquo;t need GPT-5.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;&lt;a href="https://github.com/jayluxferro/local-splitter"&gt;Local-Splitter&lt;/a&gt;&lt;/strong&gt; is an open-source shim that sits between your coding agent and the cloud. A 3B parameter model running locally on Ollama triages every request: trivial ones get answered locally (zero cloud tokens), and complex ones get their prompts compressed before forwarding. The paper is now on arXiv: &lt;a href="https://arxiv.org/abs/2604.12301"&gt;2604.12301&lt;/a&gt;.&lt;/p&gt;</description></item><item><title>Resilient Write: Giving Coding Agents a Write Path That Doesn't Break</title><link>https://sperixlabs.org/post/2026/04/resilient-write-giving-coding-agents-a-write-path-that-doesnt-break/</link><pubDate>Sun, 12 Apr 2026 22:00:00 +0000</pubDate><guid>https://sperixlabs.org/post/2026/04/resilient-write-giving-coding-agents-a-write-path-that-doesnt-break/</guid><description>&lt;p&gt;If you&amp;rsquo;ve spent any time watching an LLM coding agent work, you&amp;rsquo;ve seen it happen: the agent generates a perfectly good file, calls &lt;code&gt;Write&lt;/code&gt;, and&amp;hellip; nothing. The content vanishes. The agent retries the exact same payload. Five times. Then it gives up or cobbles together a &lt;code&gt;cat &amp;gt;&amp;gt; file.tex &amp;lt;&amp;lt;EOF&lt;/code&gt; workaround in the shell.&lt;/p&gt;
&lt;p&gt;This happened to me in April 2026 while an agent was producing a &lt;a href="https://sperixlabs.org/post/2026/04/what-leaves-your-workstation-when-you-use-an-llm-coding-cli/"&gt;telemetry report&lt;/a&gt;. A LaTeX document containing redacted HTTP headers like &lt;code&gt;Authorization: Bearer sk-ant-oat01-{REDACTED}&lt;/code&gt; got silently rejected by the host tool&amp;rsquo;s content filter. The prefix &lt;code&gt;sk-ant-&lt;/code&gt; was enough to trigger the regex. No error. No feedback. Just silence and wasted tokens.&lt;/p&gt;</description></item></channel></rss>