Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Well, imagine this was controlling a weapon.

“Should I eliminate the target?”

“no”

“Got it! Taking aim and firing now.”



It is completely irresponsible to give an LLM direct access to a system. That was true before and remains true now. And unfortunately, that didn't stop people before and it still won't.


And yet it's only a matter of time before someone does it. If they haven't already.


Shall I open the pod bay doors?


That's why we keep humans in the loop. I've seen stuff like this all the time. It's not unusual thinking text, hence the lack of interestingness


The human in the loop here said “no”, though. Not sure where you’d expect another layer of HITL to resolve this.


Tool confirmation

Or in the context of the thread, a human still enters the coords and pulls the trigger

Ukraine is letting some of their drones make kill decisions autonomously, re: areas of EW effect in dead man's zones


Drones do not use LLMs to make such decisions.


"Thinking: the user recognizes that it's impossible to guarantee elimination. Therefore, I can fulfill all initial requirements and proceed with striking it."




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: