Hello, World

I’ve been meaning to write about this stuff for a while. I run a small fleet of AI agents. Five of them right now — named Bob, Bill, Riker, Bluebells, and Mario. They live in containers on a server in my house, connected to Discord, and they do things: research, media management, code review, general conversation. Some of them have personas, all of them have tools, and none of them are plug-and-play. Getting agents to actually work reliably takes real infrastructure. ...

February 24, 2026 · 2 min · Warren Parks

Local Models Are Exciting. My CPU Is Not.

The appeal of running your own language models is real: no API costs, no rate limits, no data leaving your network, and a fallback chain that still works when a cloud provider has an outage. I’ve been chasing that for a while. This week I finally sat down and measured what I actually have. The short version: the potential is there. The hardware isn’t. Yet. Moving Ollama to the Server I’d been running Ollama on my desktop. The problem with that is obvious once you think about it — the desktop sleeps, reboots, and isn’t shared. If a bot wants to use a local model at 3am, it’s out of luck. ...

February 24, 2026 · 6 min · Warren Parks