LLMs will not be doing that. I wish they could, but they just spit out whatever without verifying anything. Even in Cursor which has the agent tell you to run the test script they generated to verify the output, it just says “yep seems fine to me!”.
AI at the current state in my workflow is a decent search engine and stackoverflow. But it has far greater pitfalls as OP pointed out (it just assumes the code is always 100% accurate and will “fake” API).
AI at the current state in my workflow is a decent search engine and stackoverflow. But it has far greater pitfalls as OP pointed out (it just assumes the code is always 100% accurate and will “fake” API).