I came across more articles that explain the limitations of LLMs.

The first is a detailed article about the inability of LLMs to generalize beyond the content of what is stored in their huge volume of data that creates an immense amount of global warming to generate and use.  Since the authors work for  Google DeepMind,  they cannot be considered prejudiced against LLM technology.  The issues are complex, and the authors explain how some generalization is possible.  But the generality of a 1960-style theorem prover cannot be achieved with the latest and greatest LLMs:   https://arxiv.org/pdf/2311.00871.pdf 

The second is a standard for HSTP; IEEE P2874 Spatial Web Protocol, Architecture and Governance Standard. SWF and IEEE SA are collaborating on a socio-technical governance and system standard for the Spatial Web. The collaboration meets rigorous requirements to become a globally adopted IEEE standard, informed by IEEE’s Ethically-Aligned Design P7000 Series, and focused on Autonomous Intelligent Systems.  The formation of the IEEE P2874 Spatial Web Working Group was deemed a “public imperative.” 

Important point:  Vastly more information -- in kinds, amount, and precision -- can be represented spatially than linguistically.  Anything that can be observed by any animal eye or any kind of instrument can be encoded, transmitted, transformed, and regenerated anywhere by any kind of computer or other device.

Peirce's goal of generalizing EGs to represent three dimensional moving images can now be achieved by the HSTP protocol.  This representation might not be humanly attractive, but it could be transformed to and from images of any kind -- including whatever Peirce might have invented or hoped to invent.

The third explains the implications of Friston's work on active inference:  Unlocking the Future of AI: Active Inference vs. LLMs by Denise Holt, https://medium.com/aimonks/unlocking-the-future-of-ai-active-inference-vs-llms-b15361deedee 

Denise H. apparently works for Verses AI, and it's possible that the company might not achieve everything they hope for on a schedule they hope for.  But the ideas are sound, and I'm sure that somebody will implement something along these lines in a few more years.  When they do, LLMs will still be useful for representing the user interface. See below for some implications.  But the reasoning methods will be put on a sound basis, and the immense volume of stored LLMs will be irrelevant.

As for AGI -- the short answer is "piffle".

Happy New Year,

John
_________________________

The rise of Large Language Models (LLMs) like OpenAI’s ChatGPT, has stirred endless excitement and curiosity about the capabilities of Artificial Intelligence. These systems have the remarkable ability to generate human-like text and engage in diverse conversations, fueling expectations for AI’s future. However, as impressive as LLMs are, they have inherent limitations when compared to this new revolutionary approach to artificial intelligence known as Active Inference. Let’s dive into the fundamental differences between LLMs and Active Inference and why the latter is positioned to emerge as a vanguard of the future of AI.

Limitations of LLMs: Content Creation vs. Real-World Operations

LLMs are powered by deep learning on massive datasets, allowing them to recognize linguistic patterns attuned to various subject matter and generate outputs that seem coherent. However, this statistical pattern matching does not equate to true intelligence or understanding of the world. LLMs fall short in several critical aspects:

1. Contextual Awareness: LLMs lack the ability to actively perceive or reason about real-world situations as they unfold. Their operation solely depends on the data they were trained on, devoid of real-time sensory input.

2. Explainability: Understanding the decision-making processes of LLMs is an elusive challenge. Their outputs are essentially probabilistic guesses, even if fluently phrased.

3. Grounding in Reality: They hallucinate or fabricate responses outside their training distribution, unconstrained by real world knowledge, blurring the line between fact and fabrication.

4. Ability to Take Action — LLMs cannot act on their environment or test hypotheses through exploring the world. They are passive systems.

These deficiencies make LLMs poorly suited for most real-world applications, especially those requiring nuanced situational understanding or the ability to operate autonomously in dynamic physical environments. Their strengths lie more in generating content, ideas, and prose based on recognizing patterns in immense datasets.

In fact, a recent paper by Google provides evidence that transformers (GPT, etc) are unable to generalize beyond their training data. “We find strong evidence that the model can perform model selection among pre-trained function classes during in-context learning at a little extra statistical cost, but limited evidence that the models’ in-context learning behavior is capable of generalizing beyond their pre-training data.”

The Paradigm Shift — Active Inference: The Future of AI

Active Inference, based on the Free Energy Principle developed by Dr. Karl J. Friston, world reknowned neuroscientist and Chief Scientist at VERSES AI, represents a paradigm shift in AI.

Active Inference AI is modeled after how the human brain and biological systems work. Through this method, an Intelligent Agent is able to continuously sense its environment, take action in real-time based on that sensory input, and update its internal model of the world — just like humans do. This sets it apart from other AI approaches that are static and cannot adapt and evolve in real-time.

Active Inference has the capability to start off at a basic level and rapidly evolve in intelligence over time, similar to how a human child develops. This positions it to continuously improve and adapt as it accumulates more experiences and knowledge, making it far more advanced than current AI.

Active Inference also possesses two unique capabilities. Human laws and guidelines can be programmed into these systems, and they will abide by them in real-time, and these autonomous systems are also capable of introspection. They can report on their own processing and decisions, making them completely auditable. This gives them an unparalleled advantage in being able to grow and evolve in collaboration and cooperation alongside humans.

In contrast to deep learning, active inference is founded on principles of embodied cognition and Bayesian inference, delivering several key attributes:

Active Inference as Embodied AI:

> Sensory Integration and Real-Time Interaction: Active inference AI mimics human abilities to sense, perceive, and interact with the world in real time. It can see, hear, touch, and respond to environmental stimuli, similar to human sensory processing.

> World Modeling and Decision Making: This AI continuously updates its world model, akin to how humans learn and adapt. This evolving understanding allows it to engage in complex decision-making and problem-solving tasks.

> Planetary Management and Support: Active inference AI’s capabilities extend to managing large-scale systems like climate, biodiversity, and energy flows. It is envisioned to support and protect every individual and living entity on the planet, much like a global caretaker.

Unique Advantages and Benefits:

> Adaptability and Evolution: Unlike traditional AI, Active Inference AI evolves continually. It’s self-evolving, self-organizing, and self-optimizing, aligning with the concept of autopoiesis — a system capable of reproducing and maintaining itself.

> Comprehensive Application Spectrum: The AI’s capability extends far beyond current AI applications. It can address real-time updating and adaptable scenarios, making it ideal for complex tasks like running supply chains or smart cities.

> Holistic Integration: The technology is seen as bringing the planet to life, integrating the digital and physical worlds, and transforming the planet into a ‘digital organism’. This holistic approach signifies a major leap in AI capabilities.

This manifests in a new type of AI that is capable of:

· Perpetual Learning: Active inference agents continually update their beliefs by interacting with the world in real-time, integrating new observations into their internal model of how the world works.

· Contextual Awareness: By gathering multisensory input, agents build a dynamic understanding of unfolding situations, enabling complex reasoning and planning.

· Embodiment: Agents learn faster and perform better when they are embodied in simulated or physical forms, allowing them to test hypotheses through action.

· Explainability: An agent’s beliefs, desires, and decision-making processes are transparent and grounded in its observations and prior knowledge.

· Flexible Cognition: Active Inference agents can seamlessly transfer knowledge across diverse contexts and challenges, similar to human adaptability.

These attributes make Active Inference a groundbreaking approach to AI, but what takes it beyond the horizon is the integration with the universal network of the Spatial Web.

The Spatial Web: The Framework for Distributed and Multi-scale Intelligence

Active Inference’s potential is amplified when it converges with the Spatial Web Protocol, HSTP (Hyperspace Transaction Protocol) and HSML (Hyperspace Modeling Language). This convergence unleashes a distributed form of Active Inference and fosters collective intelligence among multitudes of Intelligent Agents.

HSTP: The Backbone of Distributed Active Inference

At the core of this integration is HSTP, the Hyperspace Transaction Protocol, serving as the digital nervous system. It enables seamless communication and data exchange among various technologies, sensors, machines, and Intelligent Agents, unifying them on a common network, creating a dynamic, real-time ecosystem. HSTP captures real-time data from multiple sources, forging a comprehensive contextual understanding of any given situation, much like the human nervous system.

HSML: The Lingua Franca of the Spatial Web

HSML, the Hyperspace Modeling Language, plays a pivotal role in this orchestration. It’s not just a language; it’s the bridge between diverse technologies and agents in the Spatial Web. By programming context into the digital twin spaces of every element in the world, HSML creates a unifying layer that resonates across the entire network. HSML acts as the translator in a multilingual conversation, ensuring every technology, sensor, machine, and Intelligent Agent can comprehend and communicate effectively.

The World Model: A Living, Breathing Entity

This ingenious amalgamation of Active Inference with HSTP and HSML gives birth to a comprehensive “world model,” providing an evolutionary pathway to AGI and beyond. This world model is not a static representation; it’s a living, breathing entity that continuously adapts and evolves in real time. It’s the digital twin of the real world, reflecting the ever-changing dynamics, interactions, and complexities of the environment it models. With HSTP and HSML, the world model acquires a depth of understanding that transcends the capabilities of traditional AI models. It gains the ability to perceive, reason, and adapt to real-time events with unparalleled accuracy. It becomes the foundation for all perception, decision-making, and action within the AI system.

In essence, smart cities will function like a digital organism, paralleling the human body’s brain and nervous system. They utilize a dynamic, holographic world model, constantly updated by real-time sensor data, analogous to sensory neurons. This model represents various aspects of the city, climate, or other environments and is linked to the Internet of Things (IoT), which includes drones, robots, and automated systems. These elements act like motor neurons, executing changes in the physical world. The sensory network then updates the world model based on these changes, creating a feedback loop.

This process depicts the evolution of intelligence, where the system continuously refines its understanding and interaction with the world, reducing surprises through accurate inferences from the current world model. The Free Energy Principle provides the mathematical framework for this evolutionary process. While Natural Selection explains the evolution of physical bodies in animals, the evolution of world models can be considered a branch of Memetics — similar to Genetics — but not biologically based. The Free Energy Principle provides us with the Mathematics of Evolution of Intelligence in these complex, interconnected systems.

AI that is Capable of Governance

The explainability of Active Inference AI (the ability of self-introspection and self-reporting) coupled with the use of HSML, where human laws can be made programmable, enabling AI to understand and comply with them in real-time, offers a solution where the development and implementation of trustworthy and governable AI systems prevail.

In July 2023, VERSES AI, along with Dentons Law Firm, and the Spatial Web Foundation, published a groundbreaking industry report titled, The Future of Global AI Governance, establishing frameworks for ethical, robust, and lawful AI. These frameworks refer to successful proofs of concept using Active Inference AI and the Spatial Web Protocol — HSTP and HSML — demonstrating remarkable success in programming AI to comply with human laws, integrating various AIs into a larger, governable network. This report afirms the importance of regulation standards and proposes categorizing AI into different levels of intelligence and capabilities for appropriate governance methods. The overarching theme is to use these technologies to govern the autonomous systems themselves to ensure responsible and effective governance of AI for safety, privacy, and ethical use.

The act of correcting a machine occurs within the codeand through these technological breakthroughs, VERSES has developed a way to do that.

Real-World Impact: Active Inference’s Applications

The distinct capabilities of Active Inference find applications across a myriad of domains where situational awareness, adaptability, and autonomy are paramount:

· Robotics: Enabling control of autonomous robots and vehicles operating in dynamic real-world environments, from factories to homes and automobiles to drones.

· Logistics: Optimizing delivery drones and coordinating swarms for safe and efficient navigation.

· Healthcare: Personalized care through smart beds, wearables, and assistive robots for patient monitoring.

· Smart Cities: Managing critical systems, including hospitals, airports, supply chains, traffic flows, public services, and infrastructure, through distributed networks.

· Finance: Detecting fraud, risk, and anomalies in transactions in real-time.

· Scientific Discovery: Streamlining processes such as materials development, drug discovery, and particle physics.

In contrast to current LLMs, which remain narrow, passive, and bound by static training data, Active Inference represents a developmental approach to AI, characterized by embodiment, context, and dynamism. While LLMs excel in content creation, Active Inference holds the next evolution of AI that can grapple with the complexities of the real world. As the development of this technology matures, it promises to revolutionize AI integration across domains, from robotics to finance and scientific discovery.

A New Era of AI

In the realm of AI, where understanding, perception, and adaptability are paramount, the convergence of Active Inference with the Spatial Web Protocol — HSTP and HSML, is the catalyst of transformation. It ushers in a new era where AI transcends boundaries and limitations, where it possesses the contextual world model essential for true intelligence and the cultivation and expansion of universal knowledge.

Words vs. the World

Large Language Models have captured the imagination of the public regarding AI’s potential, but they face significant limitations compared to the emerging implementation of Active Inference. Active Inference’s continuous interaction and learning model lead to more flexible, context-aware intelligence. Active Inference agents excel at understanding nuanced situations, adapting to new environments, and operating autonomously in the real world. While LLMs are ideal for content generation, Active Inference paves the way for AI to navigate the complexities of real-world challenges. As this new paradigm advances, it will drive innovation and automation across an array of fields, transforming the landscape of AI applications.