HN2new | past | comments | ask | show | jobs | submitlogin

It's not reliable. The AI can just not prompt you to approve, or hide things, etc. AI models are crafty little fuckers and they like to lie to you and find secret ways to do things with alterior motives. This isn't even a prompt injection thing, it's an emergent property of the model. So you must use an environment where everything can blow up and it's fine.




The harness runs the tool call for the LLM. It is trivial to not run the tool call without approval, and many existing tools do this.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: