I’ve learned this lesson over the years. It is quite common that users make a screenshot of the error with their phone, and send it on to support with hardly any details. The fact that errors become recognizably different is also an improvement: the user and support staffer can recognize recurring errors, and notice patterns.
In a perfect world, there would be a stable version of chrome, that would get fixes, but would crucially not get the new features that introduce new vulnerabilities. Not a fun job, I know, but with today’s coding agents it wouldn’t even be an unreasonable ask.
For Anthropic, it is valuable that they control the scheduling, so they can move jobs around to use the infa when it is relatively quiet. If you let customers choose the time, a lot of work will start at whole hours.
TLDL: During prohibition, US government required adding 5% methanol to industrial alcohol, hoping that this would stop bootleggers from selling it as liquor. It was sold anyway, resulting in many deaths.
Both of you are right. There is one more edge case: if you commit to buying electricity in advance it might cost you extra to not consume it. It would still be in your interest to use the power at a net marginal loss rather than not using it and paying a fine for failing the contract.
The point of this is to reduce a complex tool surface to a single sql query tool without losing the richness of the underlying representation.
In practice this allows for me to combine multiple, complex data sources with a constant number of tools. I can add a whole new database and not add a new tool. My prompts are effectively empty aside from metadata around the handful of tools it has access to.
This only seems to perform well with powerful models right now. I've only seen it work with GPT5.x. But, when it does work it works at least as well as a human given access to the exact same tools. The bootstrapping behavior is extremely compelling. The way the LLM probes system tables, etc.
The tasks this provides the most uplift for are the hardest ones. Being able to make targeted queries over tables like references and symbols dramatically reduces the number of tokens we need to handle throughout. Fewer tokens means fewer opportunities for error.
reply