Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It’s not “just like”, because humans can be held accountable. Also, I suspect that the distribution of failure modes is actually substantially different between LLMs and humans.


Late to reply, but my point was that even with the ability to be held accountable on pain of unemployment, imprisonment or even death, sometimes even these disincentives are not enough to stop a certain small percentage of humans from going haywire. It's all just something we have to live with.


Add more humans and LLMs to correct for errors. If humans sometimes go crazy and try to randomly end the world at a rate of 0.1%, requiring two humans to turn two keys synchronously to end the world reduces the error rate to 0.01%.

So, to avoid depressed AIs ending the world randomly, have a stable of multiple AIs with different provenance (one from Anthropic, one from OpenAI, one from Google...) require a majority agreement to reduce the error rate. Adjust thresholds depending on criticality of the task at hand.


Let's hope they're not correlated.




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: