Latest Posts

When Users Won’t Wait: Engineering Killable LLM Responses

In our application, the chatbot can’t hide behind a loading spinner; users keep talking and expect it to pivot instantly. This constraint forced us to develop some lightweight techniques you can graft onto your own LLM app that serves impatient users.

In-memory free-text search is a super-power for LLMs

While working on LLM-driven NPCs, I observed significant improvements in several areas by adding a simple component: in-memory free-text search

Get the hell out of the LLM as soon as possible

Don’t let an LLM make decisions or implement business logic: they suck at that.

Four bad definitions of "Agentic AI"

If your team promises to deliver (or buy!) 'Agentic AI', then everyone needs to have a shared understanding of what that means; you don't want to be the one left trying to explain the mismatch to stakeholders six months later. There's no current (2025-03-30) widely accepted definition, so if you're using the term, be clear on what you mean, and if someone else is using the term, it's worth figuring out which one they mean.

Street-fighting RAG: Chain-of-thought prompting

or, reducing hallucination and making in-generation adjustments to LLM responses

Get these articles sent to you

If you liked it, you might like other stuff I write

Clicky