Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

One large model is not how the brain works. It’s not how org charts work.

That LLMs are capable of what they are at the compute density they are strongly signals to me that the task of making a productive knowledge worker is in overhang territory.

The missing piece isn’t LLM advancement, it’s LLM management.

Building trust in an inwardly-adversarial LLM org chart that reports to you.



The way these systems work feel massively inefficient.

We don't re-evaluate our astrophysics models when reading a cooking book.


Neither does GPT-4 or other sparse mixtures of experts, such as e.g. switch transformers [1].

[1] https://arxiv.org/abs/2101.03961


Speak for yourself :)




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: