Reflecting on some of the toughest challenges I've tackled in my career.
A capable model with ten tools burns through its step budget before it can answer anything. So Stir routes every query through a tiny classifier first, scopes the main model to a four-tool subset per intent, and short-circuits one-word queries with zero LLM calls.