Latest Posts
When Users Won’t Wait: Engineering Killable LLM Responses
In our application, the chatbot can’t hide behind a loading spinner; users keep talking and expect it to pivot instantly. This constraint forced us to develop some lightweight techniques you can graft onto your own LLM app that serves impatient users.
Read Post
In-memory free-text search is a super-power for LLMs
While working on LLM-driven NPCs, I observed significant improvements in several areas by adding a simple component: in-memory free-text search
Read Post
Get the hell out of the LLM as soon as possible
Don’t let an LLM make decisions or implement business logic: they suck at that.
Read Post
Four bad definitions of "Agentic AI"
If your team promises to deliver (or buy!) 'Agentic AI', then everyone needs to have a shared understanding of what that means; you don't want to be the one left trying to explain the mismatch to stakeholders six months later. There's no current (2025-03-30) widely accepted definition, so if you're using the term, be clear on what you mean, and if someone else is using the term, it's worth figuring out which one they mean.
Read Post
Street-fighting RAG: Chain-of-thought prompting
or, reducing hallucination and making in-generation adjustments to LLM responses
Read Post
Get these articles sent to you
If you liked it, you might like other stuff I write