I find it amusingly ironic how one comment under yours is pointing out that there’s a mistake in the model output, and the other comment under yours trusts that it’s correct but says that it isn’t “real reasoning” anyways because it knows the algorithm. There’s probably something about moving goalposts to be said here
If both criterion A and B need to be satisfied for something to be true, it’s not moving the goalposts for one person to point out A is not true, and another person to point out that B is not true.
What will really bake your noodle is when you realize that just because the model's answer is wrong doesn't mean it didn't use reasoning to reach it.
Is your reasoning always perfect? No? Ever get partial credit on a test question in school? Yes? Well, maybe don't expect perfection from a model that didn't exist 5 years ago, that was considered impossible 10 years ago, and that would have gotten you burned as a witch 15 years ago.
Nobody claims that o3-pro is AGI, or even that it is going to lead up to AGI.
Being able to manually write out hundreds of steps of the Towers of Hanoi problem is not a requirement for AGI, in much the same way that being able to manually multiply 50 digit numbers is not a requirement to be a successful mathematician.