| |

Knowledge versus Language

The biggest limitation of today’s Large Language models (LLMs)—or just ChatGPT for many people who aren’t in the AI space—is the tight coupling between the language, knowledge and context. I call today’s kind of statistical approximation ‘lossy’ because its representation cannot be converted back to the original meaning at all times.

Now I’m biased because I have solved how brains work sufficiently to vastly improve AI with an implementation that will revolutionize how our devices work. With sufficient funding we can demonstrate our differences while progressing a valuable roadmap for society. I laid this out in my book ‘How to Solve AI with Our Brain’ including my 5 year vision and how this will be achieved, in Chapters 18-20.

selective focus phot of artificial human skull

The final frontier in science: how our brain works. Imagine replicating human language correctly and talking to your devices as you would another person. What’s stopping us? Photo by jesse orrico on Unsplash

Knowledge versus language: what’s the difference?

The problem with today’s “AI” is that it models language as a sequence of words (tokens). While it is true that language is normally produced as a sequence of words, the sequences are flexible across each language and also different between the world’s languages. Full flexibility comes from the meanings of words.

In this sentence: “Can you call my wife, I mean Sue, no Samantha, um, text her?” it has nothing to do with calling. It is a command about texting Samantha even though the syntax with the words used is statistically similar to ‘calling my wife.’ Sue is just an error in the text.

Spoken language has lots of ‘errors.’

In today’s AI, transformers are used to identify the most statistical match across many dimensions. This is effective for some problems, but not others. The main use-case is human ‘conversation,’ in which the resolution of context is critical but out-of-reach of the LLM. The most statistically valid context from Background knowledge (General Common Ground—GCG) is rarely the best choice in the current context (Immediate Common Ground—ICG).

Language depends on ICG extensively to provide concise communications, exploiting ‘common sense.’ A system without ICG is unlikely to pivot to one that has it, without fundamental disruption and high cost.

Therefore, the most likely breakthrough is a system not based on statistics but based on ‘context’.

What does this all mean?

Knowledge—who said what to whom, when, where and why—is different to what is said now (to who, where, when and why). Language deals with this immediate context (ICG) to resolve the meaning of each sentence, validating with questions if needed.

In the following example, the meaning of the language: “The man bought a car from the dealer for one dollar” involves a well specified transfer of goods.

Figure 1: A man causes a car to be transferred in possession from the dealer to himself. He transferred possession of one dollar to the dealer as well. Image courtesy of Pat Labs Pty Ltd.

Notice that the context was entered into the conversation on the first line. There is no other background knowledge involved in this example—it is all ICG (current context).

The meaning of the other words in the questions like ‘pay for,’ ‘receive,’ and ‘sold’ weren’t in the initial context, but effortlessly are resolved by the meanings of the initial language (a command) and the meaning of the question.

In other words, without any training (compare to the $100m+ for an LLM training run) the system can use the language to perform valuable work as is needed in essential device interactions.

Meaning from Language

The meaning in language is powerful. We use it to learn new topics. That is the purpose of education. At universities, for example, students don’t learn the language in say, chemistry, mathematics and engineering, but instead learn subjects by using their existing language.

Figure 2: The context in figure 1 (immediate common ground) can be displayed as shown here. The final line is sufficient as a representation, while the first 4 lines explicitly specify meaningful elements. Image courtesy of Pat Labs Pty Ltd.

In the battle against wasteful uses of energy that for today’s popular AI has resulted in a proposal for $500 billion USD to create new data centers and their corresponding nuclear and coal-fired power plants, the split between knowledge (scraping the world’s Internet data and ‘ignoring’ copyright) and language (setting up an appropriate dictionary) contrasts sharply.

The system I am demonstrating here runs on a laptop easily.

The base system doesn’t include a populated knowledge representation, but language itself is so powerful as to provide useful services without external knowledge (GCG).

Conclusion and Next Steps

In my author talks, I often point out that emeritus professor of computer science, Thomas Dietterich from Oregon State University highlights the need to split language from knowledge to enable the next generation of AI to eliminate hallucinations caused by that statistical coupling.

In my author talk here (click) you can see how I present that at around the 29:30 mark.

There are strong reasons for moving away from the current ChatGPT-style of AI that cannot be trusted due to hallucinations and its high cost of operation in favor of the next generation of AI as demonstrated today. Progress requires not only a wish to achieve a goal (hopium), but also the capability to get there.

Moving back to a model that is setup as normal software is an obvious next step on the path to human-level AI that starts at a low price with low energy consumption.


Do you want to read more?

If you want to read about the application of brain science to the problems of AI, you can read my latest book, “How to Solve AI with Our Brain: The Final Frontier in Science” to explain the facets of brain science we can apply and why the best analogy today is the brain as a pattern-matcher. The book link is here on Amazon in the US.

In the cover design below, you can see the human brain incorporating its senses, such as the eyes. The brain’s use is being applied to a human-like robot who is being improved with brain science towards full human emulation in looks and capability.

Similar Posts

  • |

    5 Principles of Accelerated Language Acquisition

    Most language courses start with a textbook, a vocabulary list, and a grammar table. They’re built around what’s easy to teach, not around how the brain actually learns. After more than forty years of working at the intersection of psychology, linguistics, and learning design — and after learning Mandarin Chinese myself in six months — I’ve identified five principles that underlie rapid language acquisition. These aren’t tips or tricks. They’re the foundations. Get these right, and everything else accelerates. Get them wrong, and no amount of study hours will save you. I laid these out in my TEDx talk, “How to Learn Any Language in 6 Months.” Here, I’ll go deeper into each one, and explain why they matter whether you’re learning Mandarin, Japanese, Spanish,…

  • | |

    Understanding isn’t just memorization

    We learn all the time, continuously, regardless of our age. We never stop, but would it surprise you that many scientists propose the model where we stop learning while we are young? That is false, although more research would help prove the point. We just need some people to use experimental science!   Inside view of the ground floor of a Starbucks in Tokyo – wow! The perfect venue to learn more about language and our ability to understand — with coffee!   OK, how can I claim that learning doesn’t stop while we are young? Why so confident? Do you know the (made up) word Preada? It’s a brand that sells glasses, like Prada. Preada puts additional effort (E) into the designs of Prada,…

  • |

    Patom Theory Understands the Meaning Behind Language

    For most of my life, I have pondered a question that sits at the very center of howour brain works: how do we understand language?The question isn’t how we repeat language, nor how we recognize its surfacepatterns, but how we understand its meaning in context. If you ask ten expertsabout this subject, called Natural Language Understanding or NLU, you will getten different definitions because there are many theories available in academia!These tend to have origins in the 1950s or before and can be seen as validcompetitors in the absence of working solutions.But to most people, understanding is simple. It is the moment when wordsconnect to meaning. You do not ‘predict’ meaning (a popular paradigm in today’smachine learning community): you experience it. To many of us,…

  • How to Learn Mandarin Chinese Online

    If you want to learn Mandarin Chinese as a foreign language — what’s known in Chinese as duìwài Hànyǔ jiàoxué (对外汉语教学)— there has never been more opportunity than right now. Online tools, apps, courses, and communities have made it possible to begin learning from anywhere in the world. But opportunity doesn’t automatically mean effectiveness. And for Mandarin specifically, there are challenges that most online platforms handle poorly. As someone who learned Mandarin in six months and has spent over forty years refining the methodology behind rapid language acquisition, I want to be honest about both the challenges and the solutions. The Unique Challenges of Learning Mandarin Online Tonal Complexity Mandarin has four tones (plus a neutral tone), and word with the same pronunciation will have…

  • | |

    New language model for human conversation!

    The breathtaking view from Kobe University looking over Osaka Bay – home to the RRG 2025 bi-annual conference. Linguistic Conference: RRG 2025 The linguistic conference in Kobe, Japan, has just wrapped up. Expert linguists from around the world gave English presentations of progress over 2 days in a variety of languages including: Japanese, Taiwanese, Cantonese, Breton, Vietnamese, German, Mandarin, Mexican languages, Taiwan Sign language, and a range of African languages. They all use RRG as the model of communications. Primary developer, Robert D. Van Valin, Jr., has continued work on and growing the global community since the early 1980s. What makes Van Valin’s contributions so significant in the 20th and 21st century is its adoption of a model in which the words in a language…

  • | |

    What’s missing from AI – Part 1

    A brain in hand, a robot here ponders the use of ‘contextual meaning’ to help it emulate humans. But how do we store meaning? Photo by julien Tromeur on Unsplash Background In the 1930s, the American focus on behaviourism turned the linguistics world from the science of signs (semiotics) to one aligned with one of the great scientists in history, Pāṇini, who lived perhaps as far back as the 7th century BC. The use of Pāṇini’s linguistic model by Leonard Bloomfield led to linguistics excluding meaning, such as in the influential Chomsky monograph, Syntactic Structures, published in 1957. My proposed move back to semiotics is a side effect of the highly influential work of Robert D. Van Valin, Jr., whose development of Role and Reference…

Leave a Reply

Your email address will not be published. Required fields are marked *