Hacker Newsnew | past | comments | ask | show | jobs | submit | prng2021's commentslogin

How is anyone predicting timelines for AGI when these systems can’t do basic addition of 2 arbitrary numbers with 100% accuracy?

Can you do basic addition of 2 arbitrary numbers with 100% accuracy (no tools) ? No you can't. You will make mistakes for a sufficiently large N even with pen and paper, and a very small N without. Are you no longer generally intelligent ?

No, but I can develop methods to eventually do it.

Somewhere along the line a $10000 GPU has to be equivalent to using a finger to do arithmetic in the dust.

LLMs should use tool calling (which is 100% reliable) instead of doing math internally. But in general it would be nice to be able to teach a process and have the AI execute it deterministically. In some sense, reliability between 99% and 100% is the worst because you still can't trust the output but the verification feels like wasted effort. Maybe code gen and execution will get us there.

This is the exact problem CognOS was built to solve.

  99% reliable means you still can't remove the human from the loop — because you never know which 1% you're in. The only way to actually trust output is to attach a verifiable confidence   
  signal to each response, not just hope the aggregate accuracy holds.                                                                                                                        
                                                                                                                                                                                            
  We built a local gateway that wraps every LLM output with a trust envelope: decision trace, risk score, and an explicit PASS/REFINE/ESCALATE/BLOCK classification. The point isn't to make 
  LLMs more accurate — it's to make their uncertainty legible so the human knows when to step in.

  Open source if you want to look at the architecture: github.com/base76-research-lab/operational-cognos

"reliability between 99% and 100% is the worst because you still can't trust the output"

“I think the decision was made because the people making this decision at Anthropic are well-intentioned, driven by values, and motivated by trying to make the transition to powerful AI to go well.”

Every single one of these CEOs happily pirated unimaginable amounts of copyrighted content. That directly hurt millions of real human beings. Not just the prior creators but also crushing the future potential for success of future ones.

https://www.susmangodfrey.com/wins/susman-godfrey-secures-1-...


“You never needed 1000s of engineers to build software anyway”

What is the point of even mentioning this? We live in reality. In reality, there are countless companies with thousands of engineers making each piece of software. Outside of reality, yes you can talk about a million hypothetical situations. Cherry picking rare examples like Winamp does nothing but provide an example of an exception, which yes, also exists in the real world.


IMHO, it's flamebait. Your quoted text is provably false -- e.g., MSFT Windows, AWS, etc. and appeals to idealism of lean project teams.


This was a great article. The section “Training for the next state prediction” explains a solution using subagents. If I’m understanding it correctly, we could test if that solution is directionally correct today, right? I ask a LLM a question. It comes up with a few potential responses but sends those first to other agents in a prompt with the minimum required context. Those subagents can even do this recursively a few times. Eventually the original agent collects and analyzes subagents responses and responds to me.


Any attempt at world modeling using today's LLMs needs to have a goal function for the LLM to optimize. The LLM needs to build, evaluate and update it's model of the world. Personally, the main obstacle I found is in updating the model: Data can be large and I think that LLMs aren't good at finding correlations.


Isn't that just RL with extra power-intensive steps? (An entire model chugging away in the goal function)


That's correct, but if successful you'd essentially have updated the LLM's knowledge and capabilities "on the fly".


Maybe we could run off-peak load of that nature, when power is cheaper. Call it dreaming. ;)


And I think you basically just described the OpenAI approach to building models and serving them.


Hard to tell how widespread it is but it’s not just 1 company. https://www.webpronews.com/silicon-valleys-2025-ai-boom-revi...


Thanks for repeating what the author explained.


Where did they say that?


"The model advances both the frontier coding performance of GPT‑5.2-Codex and the reasoning and professional knowledge capabilities of GPT‑5.2, together in one model, which is also 25% faster. ... With GPT‑5.3-Codex, Codex goes from an agent that can write and review code to an agent that can do nearly anything developers and professionals can do on a computer."

They're specifically saying that they're planning for an overall improvement over the general-purpose GPT 5.2.


Did they post the knowledge cutoff date somewhere


It's here: https://platform.claude.com/docs/en/about-claude/models/over...

Reliable knowledge cutoff: May 2025, training data cutoff: August 2025


This is the thread for GPT 5.3


The author gives this example of the problem and incorrect way to leverage AI:

"Sarah was relieved. She thought she could focus on high-value synthesis work. She’d take the agent’s output and refine it, add strategic insights, make it client-ready."

Then they propose a long winded solution which is essentially the same exact thing but uses the magical term "orchestrate" a few times to make it sound different.


Well, the article was written by AI, so I wouldn't expect it to make valid arguments through a long article like this.


Already the headline is classic AI shitty writing. This isn't just x, it's basically the same thing y.


In fairness to the author, I think their point was that you take _several_ agents (not just one) and find a way to have them work like a team of 20 people. In the example, Sarah is trying to do the same job she did before, just marginally better.


Yea I guess that's accurate but they also explained that AI capabilities advance every 6-12 months and managing a team of agents buys you a few years. So their proposed solution and conclusion that it keeps you safe for years makes no sense right now. Multi agent orchestration, with an agent doing the orchestrating, is all the craze nowadays.


They made half the point, in my opinion - that you should be "doing the thing that wasn't possible before" but missed the other half - that maybe the thing you should be doing is owning and creating relationships with customers yourself instead of doing it through a company... Which maybe wasn't possible before but is now.


I agree. But the article then seems to suggest, 'you be the one left standing to orchestrate'. It didn't offer much of a suggestion about the other 20 people that would be gone.

It seemed to come down to the old 'just work better , faster, cheaper' , but that is dialed up to 11 now.


I read it more as "look for the thing that was _never done_ because no one was going to hire 20 people to do it" and all the examples were pointing out how you _should not_ try to "better, faster, cheaper" AI because you will lose quickly on all those dimensions.

I realize the irony, of course, that this article is AI-generated but it provoked something close to an epiphany for me even so.


> add strategic insights

This claim has always been BS in my experience.


“existence is a requirement to have morality. That implies that the highest good are those decisions that improve the long-term survival odds of a) humanity, and b) the biosphere.”

Those are too pie in the sky statements to be of any use in answering most real world moral questions.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: