Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

You can be pretty sure the AI isn't doing a post-hoc explanation because the only writable memory it has access to is the tokens it has already output - i.e. the explanation of the joke. Everything else is reset between every token.

As long as it comes up with different jokes different times you ask it (assuming randomness in sampling) - how could it.




The problem is it can’t remember what it hasn’t written but the end result still makes sense, so there has to be some goal after parsing the initial context that the tokens are emitted towards to. This means there’s nothing stopping it from producing an explanation, it might be in there from the very start.


Not one goal though, but a set of goals. Otherwise the random sampling of tokens wouldn't result in it outputting meaningfully different jokes.

I also think it's safe to assume that the set of goals isn't fully resolved, but it's harder to "prove" that.


There's no goal. The tokens are computed one at a time from all the previous tokens.


One is orthogonal to the other.




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: