Discussion about this post

User's avatar
skybrian's avatar

Another example of "painting itself in a corner" is that if you ask it to give an answer and then a reason for it, it sometimes gets the answer wrong and then will try to justify it, somehow. That's how it tries to escape! But if you ask it to write the justification first, then the answer, then it can often avoid the trap.

(Another way to avoid the trap might be to admit a mistake, but apparently LLM's don't get much training in doing that.)

A consequence is that whenever you ask an LLM to justify its answer, there's no reason to believe that the justification is the real reason the neural network chose its answer. Plausible justifications get invented like they would for someone else's answer. If you want the justification to possibly be real (or at least used as input), you need to ask for it before asking for the answer.

Expand full comment
Matt Asher's avatar

There have been some amazing advances with this approach to AI but it absolutely cannot reason, not in any human sense of the word. I asked Chat GPT4 for the most efficient way to tile a specific rectangular space with rectangular tiles. Dozens of attempts to get it to come up with a solution failed, despite adding lots of description, asking if the solution it provided was correct (it did at least always say that its solution was wrong), and giving it hints. Sentence completion is not reasoning, even if trained on lots of sentences that embed reasoned thinking.

For now I think any tasks that are more about reasoning than language will have to be sent out to plugins with domain specific knowledge, or actual feedback loops that can sanity check output. Though I should note that Wolfram Alpha didn't even attempt to answer the tiling question, and just gave me a definition of the first keyword in the question, like "floor" or "space".

Expand full comment
4 more comments...

No posts