I don’t understand the point you’re trying to make. LLMs are not humans.
From my perspective, the whole problem with LLMs (at least for writing code) is that it shouldn’t assume anything, follow the instructions faithfully, and ask the user for clarification if there is ambiguity in the request.
I find it extremely annoying when the model pushes back / disagrees, instead of asking for clarification. For this reason, I’m not a big fan of Sonnet 4.5.
Full instruction following looks like monkey’s paw/malicious compliance. A good way to eliminate a bug from a codebase is to delete the codebase, that type of thing. You want the model to have enough creative freedom to solve the problem otherwise you are just coding using an imprecise language spec.
I know what you mean: a lot of my prompts include “never use em-dashes” but all models forget this sooner or later. But in other circumstances I do want it to push back on something I am asking. “I can implement what you are asking but I just want to confirm that you are ok with this feature introducing an SQL injection attack into this API endpoint”
Agreed. With Claude Code I will often specify the feature I want to develop, then tell it to summarize the plan for me, give me its opinion on the plan, and ask questions before it does anything. This works very well. Often times it actually catches some piece I didn’t consider and this almost always results in usable code or code that is close enough that Claude can fix after I review what it did and point out problems.
Sometimes pushback is appropriate, sometimes clarification. The key thing is that one doesn't just blindly follow instructions; at least that's the thrust of it.
If I tell it to fetch the information using HTPP, I want it to ask if I meant HTTP, not go off and try to find a way to fetch the info using an old printing protocol from IBM.
> and ask the user for clarification if there is ambiguity in the request.
You'd just be endlessly talking to the chatbots. Humans are really bad at expressing ourselves precisely, which is why we have formal languages that preclude ambiguity.
I think the opposite. I don't want to write down everything and I like when my agents take some initiative or come up with solutions I didn't think of.
No. And for the same reason that pure "instruction following" in humans is considered a form of protest/sabotage.
https://en.wikipedia.org/wiki/Work-to-rule