• Communist@lemmy.frozeninferno.xyz
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    17 hours ago

    You’re probablistic, you just have an internal verifier, you think things that are silly, and then decide not to say them all the time. A human being often thinks things that they realize are silly before they say them… that’s an entirely unfair goal in the first place from my perspective, why does it have to be non-probablistic?

    Are you not a general intelligence because sometimes your brain thinks silly things?

    o3 currently works precisely that way, by the way, it generates hundreds of possible things, and then uses something that checks if the steps actually work, before it outputs. In fact, they then reinforce it on these correct logical steps, so it becomes better at not outputting illogical answers like you said.

    it’s interesting that you said “not on the probability of the next word, but on context and rationality”

    context IS pricesely that, you know what’s likely to come next because of the context, that’s you understanding context. YOU as a human being don’t even always get this right, you must realize we are not perfect beings, we think of possibilities and choose the right one. I think we’re much better at this right now, but i don’t think that’s a fundamental difference between us and o3.

    Rationality is the internal verifier.

    Something that doesn’t require thousands of hours of training to update and instead is capable of ingesting and rationalize new information on the fly.

    Being able to do this is… exactly what arc-agi was testing. Literally the entire point of the benchmark, it can do that.

    I’ve done the test by the way, I solved it by brute forcing possible solutions in my head, then checking if they were true… did you just divine the answers instantly?