ARCHETYPE AI

0

What Is Physical AI? – Part 1

Archetype AI Team
  • November 1, 2023

Generative AI and large language models (LLMs) have opened new avenues for innovation and understanding across the digital domain. This became possible because LLMs trained on massive internet text datasets can learn the fundamental structure of language. What if AI models could also learn the fundamental structure of physical behaviors–how the physical world around us changes over space and time? What if AI could help us understand the behaviors of real world systems, objects, and people?

At Archetype AI, we believe that this understanding could help to solve humanity’s most important problems. That is why we are building a new type of AI: physical AI, the fusion of artificial intelligence with real world sensor data, enabling real time perception, understanding, and reasoning about the physical world. Our vision is to encode the entire physical world, capturing the fundamental structures and hidden patterns of physical behaviors.

We’re building the first AI foundation model that learns about the physical world directly from sensor data, with the goal of helping humanity understand the complex behavior patterns of the world around us all. We call it a Large Behavior Model, or LBM. Here’s how it works.

Getting Data From The Real World

LLMs get their data from text—the vast corpus of writing humans have added to the internet over the last several decades. But that is the extent of their knowledge.

The web captures just a small glimpse of the true nature of the physical world. To truly understand real world behaviors, an AI model must be able to interpret much more than human-friendly data like text and images. Many physical phenomena are beyond direct human perception: too fast-moving, complex, or simply outside the realm of what our biological senses can detect. For example, we can not see or hear the temperature of objects or the chemical composition of air. An AI model trained only on a human interpretation of the physical world is inherently limited and biased in its understanding.

Our foundation model learns about the physical world directly from sensor data –  any and all sensor data. Not just cameras and microphones, but temperature sensors, inertial sensors, radar, LIDAR, infrared, pressure sensors, chemical and environmental sensors, and more. We are surrounded by sensors that capture rich spatial, temporal, and material properties of the world.

Different sensors measure different physical properties and phenomena which are related and correlated. To encode the entire physical world, information from all these sensors must be encoded, fused and processed into a single representation. This fusion enables a holistic approach in perceiving and comprehending the physical world. And by allowing the model to ingest sensor data directly from the physical world, we can have it learn about the world without human biases or perceptive limitations.

The radar signals are difficult to interpret. We are applying powerful physical AI to make sense of complex signals like this one.

Just as the petabytes of writing on the web underpin LLMs’ human-like reasoning capability, this wealth of real world sensing data can enable physical AI to develop the capability to interpret and reason about the physical world. 

But sensor data isn’t text. Incorporating multimodal sensor data into an AI model isn’t trivial because different sensors produce very different raw signals: not just in the format of the data, but also in the information that the data represents. Current ML techniques require a custom processing pipeline or model for each kind of sensor. This approach makes it infeasible to fuse and interpret data across multiple sensing modalities at scale.

Our approach to solve this is to develop a single foundation model for all sensor data, enabled by a universal sensor language.

A Universal Embedding: The “Rosetta Stone” for the Physical World

At their core, LLMs break down language into vectors – mathematical representations of language components, or tokens. These vector representations enable the meaning of any sequence of tokens to be computed by performing mathematical operations on the embedding vectors.

We can do precisely the same with sensor data. By representing the sensor data as a vector in the same embedding space as language, we can compute the meaning of the sensor data using the same mathematical operations.

The key here is the single embedding space that we use for all different sensors, modalities, and language. This universal embedding space allows information to be fused from disparate real world sources and combined with human knowledge captured through language. 

Here’s how it works:

  • We build foundational encoders for general classes of sensors. The foundation encoders convert raw sensor data into compressed vector representations that retain information about the essential structure of the signal.
  • These vectors are then projected into the high-dimensional universal embedding space. Because all the sensor data is now encoded as vectors, patterns in the sensor data are represented as vector relationships in this space.
  • Text embeddings are also projected into this space, allowing sensor data to be mapped to semantic concepts.
  • Semantic information is then decoded from the universal embedding space into human-understandable formats such as natural language, APIs, or visualizations using multimodal decoders.

The universal embedding space thus acts as a “Rosetta Stone” for deciphering meaning from different kinds of physical world data. It allows natural human interpretation and interaction with physical sensor data. 

Block diagram of the Newton physical AI foundation model

By mapping disparate sensor data into a single space, it also allows us to use sensors we understand to interpret the ones we don’t. For example, we can capture radar data with camera data and use image labeling on video to label the radar data. Or we can correlate audio recordings from microphones with vibration sensors to monitor and diagnose machinery health, using the sounds and vibrations they emit to predict failures before they happen. 

Check out Part 2 of this blog post, where we will explore use cases and practical applications for physical AI in the real world.

Recommended posts

Humans can instantly connect scattered signals—a child on a bike means school drop-off; breaking glass at night signals trouble. Despite billions of sensors, smart devices haven’t matched this basic human skill.  Archetype AI’s Newton combines simple sensor data with contextual awareness to understand events in the physical world, just like humans do.  Learn how it’s transforming electronics, manufacturing, and automotive experiences.

December 12, 2024

Implementing AI in industrial settings comes with significant challenges like ensuring employee safety, estimating productivity, and monitoring hazards—all requiring real-time processing. However, sending sensor data to the cloud for analysis introduces latency and security concerns, driving up costs. The solution? Eliminate the cloud. With Archetype AI’s Newton foundation model, AI can run on local machines using a single off-the-shelf GPU, delivering low latency, high security, and reduced costs in environments like manufacturing, logistics, transportation, and construction.

September 5, 2024

At Archetype, we want to use AI to solve real world problems by empowering organizations to build for their own use cases. We aren’t building verticalized solutions –instead, we want to give engineers, developers, and companies the AI tools and platform they need to create their own solutions in the physical world.

We are excited to share a milestone in our journey toward developing a physical AI foundation model. In a recent paper by the Archetype AI team, "A Phenomenological AI Foundation Model for Physical Signals," we demonstrate how an AI foundation model can effectively encode and predict physical behaviors and processes it has never encountered before, without being explicitly taught underlying physical principles. Read on to explore our key findings.

October 17, 2024

Currently, digital companions are the dominant metaphor for understanding AI systems. However, as the field of generative AI continues to evolve, it's crucial to examine how we frame and comprehend these technologies—it will influence how we develop, interact with, and regulate AI. In this blog post, we'll explore different metaphors used in AI products and discuss how they shape our mental models of AI systems.

September 25, 2024