The replacement of humans, who do white collar work, is a constant topic of conversation in the present times. The world is awash with talk of the intelligence of LLMs, and the oncoming threat of AGI. But what are the limitations of LLMs, and what, if anything, do humans have as an advantage to LLMs.
I’ll list what I think below. Some of these may seem l
The Personal Touch
This article is written by a human*, but the image above is generated by an AI (ChatGPT 4o). The power and productivity leveraging by AI is clear-as-day; the image above would have taken me many hours of work to produce such a clean result, and AI has provided it in seconds. However, it is not my image, it is not what I pictured in my mind when I thought of the image first, but it does the same job. It is something that fits the prompt, however vague that is.
Now for someone, who is an expert artist (not me), to produce their minds-eye image, it may be easier to create the image, than to describe exactly what they want; and in fact, for any written work (code/writing etc.) it is in the very act of writing that you are describing what you want. If what you describe in a prompt, is enough to produce the correct result, then by definition the prompt is enough. The rest is just the act of translation, the presentation of the information in an easily digestible form. However, the actual information content will be the same.
Obviously, this changes if you do not know what to write, or you do not know enough to create the content in the first place. Then the AI acts as a search engine, gathering and collating information, as you would do normally on a day-to-day.
So, what does this mean. The human value is therefore, the personal touch, the uniqueness that you bring to the table, that is a result of your own experience. This manifests as your ability to describe your own inner thoughts, be it through writing, images, or any creative outlet. It is this unique perspective that has always been difficult to describe to others, and is equally difficult to describe to an AI.
Symbolic Reasoning/Abstraction
This is a complex one, and requires an understanding of what symbolic reasoning is, and how it leads to abstraction. Of the current LLMs, including “reasoning” ones, this is the major downfall, and is something (at time of writing) that is still being worked on.
According to ChatGPT 4o, symbolic reasoning is:
The process of manipulating abstract symbols according to logical rules to draw conclusions or solve problems.
This definition not only applies to mathematics, but also to the symbolism of literature. The ability to use metaphor to substitute one symbol for another in a film or book, is a way to make abstract topics relatable. For example, in horror films, primal fears, become manifest in monstrous form, and the situation and context is built around them with is analogous to reality, or how that fear comes about in reality. The ability to swap the symbol of a monster in place of fear, is a kind of symbolic manipulation, and the understanding of the context is the reasoning.
So, symbolic reasoning is key to abstract and higher level thinking. It is in some ways the embodiment of intelligence, but is missing from LLMs. To understand why, you need to understand the transformer architecture. I won’t go into a large amount of detail here (please refer to the link article), but LLMs excel at pattern matching.
Pattern matching is LLMs can be demonstrated by a counter-example, something that is difficult for them to solve (without external assistance). Let’s take a string of words, and ask LLMs to sort them:
banana, apple, cucumber, apricot
We then provide the LLM with a prompt:
Please sort the following: banana, apple, cucumber, apricot
Now the transformer uses something called self-attention, that is it looks at each word in the prompt, and asks which words are most relevant for the other words. Crucially, this attention mechanism is built into the system, and is fixed, but does change based on the words in the prompt. You can view the first four words as a kind of instruction to prep this matrix; but the matrix has been trained, not on the four words following (or more specifically on their ordering). To produce the next word in the response, the attention matrix itself would have to shift attention to the word apple; but the statistical nature of the training means this won’t be the most common case. In fact, in general, the attention should be even amongst all the words. This leads on to symbolic reasoning.
From the brief explanation above, it is clear to see that the LLM does not see symbols. In a human mind, each position in the list is a different symbol, and so we can manipulate them as such; not simply something to shift attention to. An LLM does not see it that way, and never will. This is clearly demonstrated by the ARC-AGI challenge which humans can do almost instantly. How human brains do it is a mystery, but I try to explore it here.
If anyone ever reads this, I know they will try the above prompt, and see that it does in fact work. This is because behind-the-scenes, the AI models have access to tools. They can convert the request to code, and run the code locally. It may seem like this solves the problem, but the point is that the symbolic reasoning should be embedded in the intelligence to perform true abstract reasoning.
* All but the headline, and otherwise stated, that is.

Leave a Reply