Research, standards and thoughts for the digital world

Earlier posts by categories:

MPAI MPEG ISO

Basic Knowledge: The Generalist Engine Getting Sharper with Every Prompt

  • Post author:
  • Post category:MPAI

Basic Knowledge is the core language model of the Autonomous User – the “knows-a-bit-of-everything” brain. It’s the first responder to a prompt but it doesn’t fire off just one answer but four of them in a progressive refinement loop, providing smarter and more context-aware responses with every refined prompt. We have already presented the system diagram of the Autonomous User (A-User), an autonomous agent able to move and interact (walk, converse, do things, etc.) with another User in a metaverse.…

Continue ReadingBasic Knowledge: The Generalist Engine Getting Sharper with Every Prompt

Domain Access: The Specialist Brain Plug-in for the Autonomous User

  • Post author:
  • Post category:MPAI

While the Basic Knowledge module is a generalist language model that “knows a bit of everything”, Domain Access is the expert layer that enables the Autonomous User to tap into domain-specific intelligence for deeper understanding of user utterances and operational context. We have already presented the system diagram of the Autonomous User (A-User), an autonomous agent able to move and interact (walk, converse, do things, etc.) with another User in a metaverse. The latter User may be an A-User or…

Continue ReadingDomain Access: The Specialist Brain Plug-in for the Autonomous User

Prompt Creation: Where Words Meet Context

The Prompt Creation module is the storyteller and translator in the Autonomous User’s “brain”, It takes raw sensory input  –  audio and visual spatial data of Context (as objects in a scene with their position, orientation and velocity) and the User State (rich description of the A‑User’s understanding of the “internal state” of the User) – and turns it into a well‑formed prompt that Basic Knowledge can actually understand and respond to. We have already presented the system diagram of…

Continue ReadingPrompt Creation: Where Words Meet Context

Visual Spatial Reasoning: The Vision‑Aware Interpreter

  • Post author:
  • Post category:MPAI

Autonomous User (A-User) is an autonomous agent able to move and interact (converse, etc.) with another User in a metaverse. It is a “conversation partner in a metaverse interaction” with the User, itself an A-User or an H-User directly controlled by a human. The figure shows a diagram of the A-User while the User generates audio-visual streams of information and possibly text as well. This is the fourth of a sequence of posts aiming to illustrate the architecture of an…

Continue ReadingVisual Spatial Reasoning: The Vision‑Aware Interpreter

Audio Spatial Reasoning: The Sound-Aware Interpreter

  • Post author:
  • Post category:MPAI

Autonomous User (A-User) is an autonomous agent able to move and interact (converse, etc.) with another User in a metaverse. It is a “conversation partner in a metaverse interaction” with the User, itself an A-User or and H-User directly controlled by a human. The figure shows a diagram of the A-User while the User generates audio-visual streams of information and possibly text as well. We have already presented the system diagram of the Autonomous User (A-User), an autonomous agent able…

Continue ReadingAudio Spatial Reasoning: The Sound-Aware Interpreter

Context Capture: The A-User’s First Glimpse of the World

  • Post author:
  • Post category:MPAI

Autonomous User (A-User) is an autonomous agent able to move and interact (converse, etc.) with another User in a metaverse. It is a “conversation partner in a metaverse interaction” with the User, itself an A-User or and H-User directly controlled by a human. The figure shows a diagram of the A-User while the User generates audio-visual streams of information and possibly text as well. The sequence of posts – of which this is the second – that illustrates more in…

Continue ReadingContext Capture: The A-User’s First Glimpse of the World