Discussion about this post

User's avatar
Luke Thorburn's avatar

Echoing the other comments: I don't think it is true to say that GPT4 is "optimised" for probabilistic reasoning. It's optimised for next-token prediction, and then for what responses humans like or fit with OpenAI guidelines using RLHF. Even if part of fine-tuning was on reasoning problems with reliable solutions, the underlying training on next-token prediction will still be significantly influencing the responses given by the model. There's a high chance these "biases" are just mimicking human behaviour.

I appreciate and admire all your work! But I don't think this idea holds up.

Expand full comment
Mitchell Porter's avatar

Bard had some thoughts: https://pastebin.com/2PtVAaGt

Expand full comment
7 more comments...

No posts