Discussion about this post

User's avatar
Alex Tolley's avatar

I 100% disagree with your premise. LLMs are guessing at the correct response in both correct and incorrect answers, but that is not the same as humans making experimental errors. It is more like human BS, or making illogical guesses at an answer. When we make errors in experiments, it is for a number of reasons, but guessing is not one. I don't even think it is similar to making correlation errors, or even extrapolation errors, as we can quickly check on LLM output, whereas extrapolations, almost by definition, have no data in the extrapolation regime to rely on.

Human BS is based on a lack of facts and poor logic. Politicians seem to be particularly prone to this, but probably, IRL, no more prone to this than most people, just that they are outspoken in public for all to see. Groupthink can lead in the same direction, as can following the thoughts of cult leaders. LLMs using statistical word and sentence prediction work similarly, IMO. However, when we do science or math, we don't do this. Experiments are designed with controls. Conclusions are drawn based on results. Discussions of the results, if speculative, are accepted as being beyond the data and are to be taken with skepticism. Math requires attention to the correct manipulations of symbols, not guessing at answers like young children.

Therefore, to be better, LLMs need to have different architectures. We can reduce hallucinations and improve accuracy, but I fear that unless we can add true understanding, their hallucinations may be as difficult to control as human dreaming, where we accept strange situations and actions that we would not IRL. This may require *ahem* consciousness, of some [limited?] sort. Enough to be able to work through possible guesses and strip out bad answers, and to work through logical chains without the same problems afflicting philosophers. [How does philosophy, with its attention to logic, arrive at different answers to teh same question?] The mixture of Experts models being tried in one way to improve accuracy, but I think just a palliative, rather like random forest decision trees. Understanding a mechanism to think through problems and how to solve them requires a different approach, and possibly an architecture that mimics the human thought processes of experts. When LLMs provide bogus citations, it implies that these are decorative and not actually used to extract information. Even ones that do provide real citations, they do not always make the correct determination of what the source says.

In some ways, Doug Lenat's Cyc was a theoretically better way to create an AI, but it fell apart under the weight of its many pieces of data objects. My interpretation is that AIs must have teh ability to store knowledge, algorithms for extracting and using this knowledge from sources and new data, and then apply the LLMs in the role of an interface to convey the answers. Having a BS machine, operating more like Kahneman's System 1 (fast) thinking, is not the way to go if we want AIs to be more than glib answering machines.

Expand full comment
Andrew Condon's avatar

What we have here with respect to the LLMs is something like an incommensurability problem - we, humans, do terrible violence to the facts by compressing things into narratives and the algorithms do terrible violence (is it really anything like as bad as humans? I’m not sure) by compressing everything into token prediction.

But humans for sure do the hallucination thing too all the time! And they’re mostly pretty unaware of it or play it down when it happens.

Expand full comment
16 more comments...

No posts