Disobedient and "rude"
The model is excellent in reasoning and such, but it's quite "disobedient" and even (in some sense) "rude", in regards to modifications to its instructions given during agentic coding.
For example, if you kick off a task, it can work through that. If it stops and asks something, and you respond, it can work through that appropriately (usually).
BUT, if you interrupt its reasoning, and tell it that it should do something else, or that it's going off course and needs to follow the original instruction, it often just continues on, giving little or not regard to the additional instruction.
AT LEAST, instead of being ignored, instead of output of the form "<further_output"> I would expect "OK, I hear you, I'll incorporate that as I go forward. ", or (if it must) "The user says . Actually, I think that's incorrect due to ; I'm still fully aligned with everything the user has requested, and so I can to proceed as before. In this framework, the user can simply interrupt me again if I've gotten this wrong, so I'll proceed. ".
Actually, it's disobedient from the outset, too. For example, if I tell it to make a skill, and to search online for the format of skill.md files, it will decide to initially search the filesystem for skill.md files instead. Not ideal.
To clarify: I believe there's a gap in the training dataset here: it's likely trained on a simple back and forth flow, whereas the training should include interruptions for questions, additions, course corrections (of the model's CoT), and "changes of mind" (of the user) and so on.