Alex,

The article you cited is interesting, and I recommend it as a useful technique for limiting the possible hallucinations of LLMs.  But the cartoon I copy below is just as applicable to this article as it is to any other application of LLMs.  There is no breakthrough here.

Furthermore, the authors' claim that their system understands or comprehends anything is absurd.  What their system generates is "a large pile of linear algebra" that does not differ from the pile in the cartoon in any essential way.   The only reason why their system performs better than the huge pile generated by OpenGPT is that it's restricted to peer-reviewed scientific articles.  The main reason why the results are fairly good is that different scientific disciplines use very different terminology.  Therefore, the texts from different articles do not mix or interfere or pollute one another.

Please note what the authors have done:  They created a collection of LLMs from a collection of published scientific articles from multiple disciplines and used LLMs to represent both English text and diagrams in those texts.   

Mixing the two different kinds of syntax is a useful enhancement, but there is nothing new in the underlying technology.   You can get the same or better enhancement by mixing data in three very different linear syntaxes:  English, SQL, and OWL.  IT's useful to use the same spelling  for the same concepts.  But if there are enough examples, the LLMs are able to detect the similarities and do the equivalent translations.

The information in the diagrams is expressed in the same words as the English text, but the two-dimensional syntax of the diagrams represents a second language.  There is nothing new there, since LLMs can relate languages with different syntax.  For any language processor,  the difference between a linear string and a 2-D diagram is trivial.   When you send the diagram to another system, the 2-D syntax  is mapped to a 1-D syntax that uses the same kinds of syntactic markers as describing a 2-D diagram in English.

Fundamental principle:  For the LLMs, it's irrelevant whether the source is a linear language or a system of diagrams that were mapped to a linear string.  The result is a pile of linear algebra.  See the cartoon.
 
John 

From: "Alex Shkotin" <alex.shkotin@gmail.com>

John,


Sure! Theoretical knowledge may be checked only by the theoretical knowledge handling system.

Meanwhile it may be interesting LMM++ advancement here https://arxiv.org/abs/2407.04903?fbclid=IwZXh0bgNhZW0CMTEAAR0oph0y 


They don't lose hope.   [JFS:  More precisely, they have no hope.  They are just confusing the issues.]


Alex



вс, 14 июл. 2024 г. в 21:40, John F Sowa <:" style="box-sizing: border-box; color: rgb(0, 102, 147); text-decoration: underline; user-select: auto;">sowa@bestweb.net>:

Peter,

Thanks for that link.  That cartoon is a precise characterization of how LLMs process data.  It was drawn in the 1990s when linear algebra usually meant something computed with matrices.  LLMs go one step farther by using tensors, but the results are in the same ballpark (or sewer).


Fundamental principle:  any machine learning system must be used with a system for evaluating or checking the answers.  For simple factual questions, a database can be used.  For more complex questions, logical deduction is necessary.  For any kind of system, ontology can detect obvious hallucinations, but ontology by itself is insufficient to detect incorrect details that happen to be in the correct category.

John